no code implementations • 25 Oct 2022 • Sidharth Ranjan, Marten Van Schijndel, Sumeet Agarwal, Rajakrishnan Rajkumar
By showing that different priming influences are separable from one another, our results support the hypothesis that multiple different cognitive mechanisms underlie priming.
no code implementations • 25 Oct 2022 • Sidharth Ranjan, Marten Van Schijndel, Sumeet Agarwal, Rajakrishnan Rajkumar
While prior work has shown that a number of factors (e. g., information status, dependency length, and syntactic surprisal) influence Hindi word order preferences, the role of discourse predictability is underexplored in the literature.
no code implementations • 10 Jul 2022 • Kunal Dahiya, Nilesh Gupta, Deepak Saini, Akshay Soni, Yajun Wang, Kushal Dave, Jian Jiao, Gururaj K, Prasenjit Dey, Amit Singh, Deepesh Hada, Vidit Jain, Bhawna Paliwal, Anshul Mittal, Sonu Mehta, Ramachandran Ramjee, Sumeet Agarwal, Purushottam Kar, Manik Varma
This paper identifies that memory overheads of popular negative mining techniques often force mini-batch sizes to remain small and slow training down.
no code implementations • CVPR 2022 • Anshul Mittal, Kunal Dahiya, Shreya Malani, Janani Ramaswamy, Seba Kuruvilla, Jitendra Ajmera, Keng-hao Chang, Sumeet Agarwal, Purushottam Kar, Manik Varma
On the other hand, XC methods utilize classifier architectures to offer superior accuracies than embedding-only methods but mostly focus on text-based categorization tasks.
1 code implementation • 12 Nov 2021 • Kunal Dahiya, Deepak Saini, Anshul Mittal, Ankush Shaw, Kushal Dave, Akshay Soni, Himanshu Jain, Sumeet Agarwal, Manik Varma
Scalability and accuracy are well recognized challenges in deep extreme multi-label learning where the objective is to train architectures for automatically annotating a data point with the most relevant subset of labels from an extremely large label set.
1 code implementation • 1 Aug 2021 • Anshul Mittal, Kunal Dahiya, Sheshansh Agrawal, Deepak Saini, Sumeet Agarwal, Purushottam Kar, Manik Varma
This paper develops the DECAF algorithm that addresses these challenges by learning models enriched by label metadata that jointly learn model parameters and feature representations using deep networks and offer accurate classification at the scale of millions of labels.
1 code implementation • 31 Jul 2021 • Anshul Mittal, Noveen Sachdeva, Sheshansh Agrawal, Sumeet Agarwal, Purushottam Kar, Manik Varma
This paper presents ECLARE, a scalable deep learning architecture that incorporates not only label text, but also label correlations, to offer accurate real-time predictions within a few milliseconds.
1 code implementation • SCiL 2021 • Hritik Bansal, Gantavya Bhatt, Sumeet Agarwal
However, we observe that several RNN types, including the ONLSTM which has a soft structural inductive bias, surprisingly fail to perform well on sentences without attractors when trained solely on sentences with attractors.
no code implementations • 3 Oct 2020 • Ayush Srivastava, Oshin Dutta, Prathosh AP, Sumeet Agarwal, Jigyasa Gupta
In the last few years, compression of deep neural networks has become an important strand of machine learning and computer vision research.
1 code implementation • ACL 2020 • Gantavya Bhatt, Hritik Bansal, Rishubh Singh, Sumeet Agarwal
Long short-term memory (LSTM) networks and their variants are capable of encapsulating long-range dependencies, which is evident from their performance on a variety of linguistic tasks.
Ranked #29 on
Language Modelling
on WikiText-103
(Validation perplexity metric)
no code implementations • 25 Sep 2019 • Kunal Dahiya, Anshul Mittal, Deepak Saini, Kushal Dave, Himanshu Jain, Sumeet Agarwal, Manik Varma
The objective in deep extreme multi-label learning is to jointly learn feature representations and classifiers to automatically tag data points with the most relevant subset of labels from an extremely large label set.
no code implementations • WS 2019 • Sidharth Ranjan, Sumeet Agarwal, Rajakrishnan Rajkumar
Based on the Production-Distribution-Comprehension (PDC) account of language processing, we formulate two distinct hypotheses about case marking, word order choices and processing in Hindi.
no code implementations • NAACL 2019 • Samvit Dammalapati, Rajakrishnan Rajkumar, Sumeet Agarwal
This study examines the role of three influential theories of language processing, \textit{viz.
no code implementations • 31 Mar 2019 • Aditi Jha, Sumeet Agarwal
Visual scene understanding often requires the processing of human-object interactions.
no code implementations • WS 2018 • Ayush Jain, Vishal Singh, Sidharth Ranjan, Rajakrishnan Rajkumar, Sumeet Agarwal
According to the UNIFORM INFORMATION DENSITY (UID) hypothesis (Levy and Jaeger, 2007; Jaeger, 2010), speakers tend to distribute information density across the signal uniformly while producing language.
no code implementations • 22 Sep 2017 • Abhimanyu Dubey, Sumeet Agarwal
The study of virality and information diffusion online is a topic gaining traction rapidly in the computational social sciences.
no code implementations • COLING 2016 • Ashwini Vaidya, Sumeet Agarwal, Martha Palmer
To build our system, we carry out a linguistic analysis of Hindi LVCs using Hindi Treebank annotations and propose two new features that are aimed at capturing the diversity of Hindi LVCs in the corpus.
no code implementations • 12 Sep 2016 • Abhimanyu Dubey, Jayadeva, Sumeet Agarwal
We compare several ConvNets with different depth and regularization techniques with multi-unit macaque IT cortex recordings and assess the impact of the same on representational similarity with the primate visual cortex.