1 code implementation • EMNLP 2021 • Tejas Indulal Dhamecha, Rudra Murthy V, Samarth Bharadwaj, Karthik Sankaranarayanan, Pushpak Bhattacharyya
We hypothesize and validate that multilingual fine-tuning of pre-trained language models can yield better performance on downstream NLP applications, compared to models fine-tuned on individual languages.
Multiple Choice Question Answering (MCQA)
Natural Language Inference
+2
1 code implementation • EMNLP 2021 • Saneem Ahmed Chemmengath, Vishwajeet Kumar, Samarth Bharadwaj, Jaydeep Sen, Mustafa Canim, Soumen Chakrabarti, Alfio Gliozzo, Karthik Sankaranarayanan
Weakly-supervised table question-answering(TableQA) models have achieved state-of-art performance by using pre-trained BERT transformer to jointly encoding a question and a table to produce structured query for the question.
1 code implementation • 29 Jun 2021 • Ashish Mittal, Samarth Bharadwaj, Shreya Khare, Saneem Chemmengath, Karthik Sankaranarayanan, Brian Kingsbury
Spoken intent detection has become a popular approach to interface with various smart devices with ease.
1 code implementation • NAACL (ACL) 2022 • Yannis Katsis, Saneem Chemmengath, Vishwajeet Kumar, Samarth Bharadwaj, Mustafa Canim, Michael Glass, Alfio Gliozzo, Feifei Pan, Jaydeep Sen, Karthik Sankaranarayanan, Soumen Chakrabarti
Recent advances in transformers have enabled Table Question Answering (Table QA) systems to achieve high accuracy and SOTA results on open domain datasets like WikiTableQuestions and WikiSQL.
1 code implementation • 1 Apr 2021 • Anuj Diwan, Rakesh Vaideeswaran, Sanket Shah, Ankita Singh, Srinivasa Raghavan, Shreya Khare, Vinit Unni, Saurabh Vyas, Akash Rajpuria, Chiranjeevi Yarra, Ashish Mittal, Prasanta Kumar Ghosh, Preethi Jyothi, Kalika Bali, Vivek Seshadri, Sunayana Sitaram, Samarth Bharadwaj, Jai Nanavati, Raoul Nanavati, Karthik Sankaranarayanan, Tejaswi Seeram, Basil Abraham
For this purpose, we provide a total of ~600 hours of transcribed speech data, comprising train and test sets, in these languages including two code-switched language pairs, Hindi-English and Bengali-English.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • COLING 2020 • Jaydeep Sen, Tanaya Babtiwale, Kanishk Saxena, Yash Butala, Sumit Bhatia, Karthik Sankaranarayanan
We posit that deploying ontology reasoning over domain semantics can help in achieving better natural language understanding for QA systems.
no code implementations • 7 Nov 2020 • Abhijit Mishra, Md Faisal Mahbub Chowdhury, Sagar Manohar, Dan Gutfreund, Karthik Sankaranarayanan
The model takes as input a set of un-ordered keywords, and part-of-speech (POS) based template instructions.
1 code implementation • 2 Mar 2020 • Saneem Ahmed Chemmengath, Soumava Paul, Samarth Bharadwaj, Suranjana Samanta, Karthik Sankaranarayanan
Zero-shot learning (ZSL) algorithms typically work by exploiting attribute correlations to be able to make predictions in unseen classes.
no code implementations • IJCNLP 2019 • Abhijit Mishra, Tarun Tater, Karthik Sankaranarayanan
In this paper, we propose a novel framework for sarcasm generation; the system takes a literal negative opinion as input and translates it into a sarcastic version.
no code implementations • ACL 2019 • Abhijit Mishra, Anirban Laha, Karthik Sankaranarayanan, Parag Jain, Saravanan Krishnan
In this tutorial, we wish to cover the foundational, methodological, and system development aspects of translating structured data (such as data in tabular form) and knowledge bases (such as knowledge graphs) into natural language.
1 code implementation • ACL 2019 • Priyanka Agrawal, Parag Jain, Ayushi Dalmia, Abhishek Bansal, Ashish Mittal, Karthik Sankaranarayanan
Semantic parsing over multiple knowledge bases enables a parser to exploit structural similarities of programs across the multiple domains.
no code implementations • ICLR 2019 • Saneem Ahmed Chemmengath, Samarth Bharadwaj, Suranjana Samanta, Karthik Sankaranarayanan
An unintended consequence of feature sharing is the model fitting to correlated tasks within the dataset, termed negative transfer.
no code implementations • TACL 2019 • Amrita Saha, Ghulam Ahmed Ansari, Abhishek Laddha, Karthik Sankaranarayanan, Soumen Chakrabarti
On one of the hardest class of programs (comparative reasoning) with 5{--}10 steps, CIPITR outperforms NSM by a factor of 89 and memory networks by 9 times.
no code implementations • NeurIPS 2018 • Anirban Laha, Saneem A. Chemmengath, Priyanka Agrawal, Mitesh M. Khapra, Karthik Sankaranarayanan, Harish G. Ramaswamy
Converting an n-dimensional vector to a probability distribution over n objects is a commonly used component in many machine learning tasks like multiclass classification, multilabel classification, attention mechanisms etc.
1 code implementation • ACL 2019 • Sai Surya, Abhijit Mishra, Anirban Laha, Parag Jain, Karthik Sankaranarayanan
The paper presents a first attempt towards unsupervised neural text simplification that relies only on unlabeled text corpora.
Ranked #2 on
Text Simplification
on ASSET
1 code implementation • CL 2019 • Anirban Laha, Parag Jain, Abhijit Mishra, Karthik Sankaranarayanan
We present a framework for generating natural language description from structured data such as tables; the problem comes under the category of data-to-text natural language generation (NLG).
1 code implementation • 10 Sep 2018 • Parag Jain, Abhijit Mishra, Amar Prakash Azad, Karthik Sankaranarayanan
We propose a novel framework for controllable natural language transformation.
no code implementations • 2 Sep 2018 • Disha Shrivastava, Abhijit Mishra, Karthik Sankaranarayanan
Our evaluation shows that the measured coherence scores are positively correlated with the ground truth for both the datasets.
1 code implementation • ACL 2018 • Amrita Saha, Rahul Aralikatte, Mitesh M. Khapra, Karthik Sankaranarayanan
We propose DuoRC, a novel dataset for Reading Comprehension (RC) that motivates several new challenges for neural approaches in language understanding beyond those offered by existing RC datasets.
2 code implementations • NAACL 2018 • Preksha Nema, Shreyas Shetty, Parag Jain, Anirban Laha, Karthik Sankaranarayanan, Mitesh M. Khapra
For example, while generating descriptions from a table, a human would attend to information at two levels: (i) the fields (macro level) and (ii) the values within the field (micro level).
1 code implementation • NAACL 2018 • Parag Jain, Anirban Laha, Karthik Sankaranarayanan, Preksha Nema, Mitesh M. Khapra, Shreyas Shetty
Structured data summarization involves generation of natural language summaries from structured input data.
1 code implementation • 31 Jan 2018 • Amrita Saha, Vardaan Pahuja, Mitesh M. Khapra, Karthik Sankaranarayanan, Sarath Chandar
Further, unlike existing large scale QA datasets which contain simple questions that can be answered from a single tuple, the questions in our dialogs require a larger subgraph of the KG.
no code implementations • 18 Jul 2017 • Disha Shrivastava, Saneem Ahmed CG, Anirban Laha, Karthik Sankaranarayanan
Our proposed learning framework is applicable to all creative domains; yet we evaluate it on a dataset of movies created from IMDb and Rotten Tomatoes due to availability of audience and critic scores, which can be used as proxy ground truth labels for creativity.
no code implementations • 18 Jul 2017 • Parag Jain, Priyanka Agrawal, Abhijit Mishra, Mohak Sukhwani, Anirban Laha, Karthik Sankaranarayanan
Existing Natural Language Generation (NLG) systems are weak AI systems and exhibit limited capabilities when language generation tasks demand higher levels of creativity, originality and brevity.
no code implementations • 1 Apr 2017 • Amrita Saha, Mitesh Khapra, Karthik Sankaranarayanan
With this dataset, we propose 5 new sub-tasks for multimodal conversations along with their evaluation methodology.