no code implementations • 14 Aug 2024 • Subhabrata Dutta, Timo Kaufmann, Goran Glavaš, Ivan Habernal, Kristian Kersting, Frauke Kreuter, Mira Mezini, Iryna Gurevych, Eyke Hüllermeier, Hinrich Schuetze
While there is a widespread belief that artificial general intelligence (AGI) -- or even superhuman AI -- is imminent, complex problems in expert domains are far from being solved.
1 code implementation • 17 May 2024 • Anwoy Chatterjee, Eshaan Tanwar, Subhabrata Dutta, Tanmoy Chakraborty
We design a cross-task prompting setup with three LLMs and show that LLMs achieve significant performance improvements despite no examples from the target task in the context.
no code implementations • 2 Apr 2024 • Gurusha Juneja, Subhabrata Dutta, Tanmoy Chakraborty
The solver model generates the solution to the subproblems that are then checked by the verifier module; depending upon the feedback from the verifier, the reasoning context is constructed using the subproblems and the solutions.
1 code implementation • 28 Feb 2024 • Subhabrata Dutta, Joykirat Singh, Soumen Chakrabarti, Tanmoy Chakraborty
Despite superior reasoning prowess demonstrated by Large Language Models (LLMs) with Chain-of-Thought (CoT) prompting, a lack of understanding prevails around the internal mechanisms of the models that facilitate CoT generation.
1 code implementation • 9 Dec 2023 • Subhabrata Dutta, Joykirat Singh, Ishan Pandey, Sunny Manchanda, Soumen Chakrabarti, Tanmoy Chakraborty
In this paper, we start with the hypothesis that much smaller LMs, which are weak at multi-step reasoning, can achieve reasonable arithmetic reasoning if arithmetic word problems are posed as a formalize-then-solve task.
Ranked #12 on Math Word Problem Solving on SVAMP (using extra training data)
1 code implementation • 21 Oct 2023 • Gurusha Juneja, Subhabrata Dutta, Soumen Chakrabarti, Sunny Manchanda, Tanmoy Chakraborty
Additionally, we show that DaSLaM is not limited by the solver's capabilities as a function of scale; e. g., solver LMs with diverse sizes give significant performance improvement with our solver-agnostic decomposition technique.
Ranked #6 on Overall - Test on JEEBench (using extra training data)
1 code implementation • 10 May 2023 • Eshaan Tanwar, Subhabrata Dutta, Manish Borthakur, Tanmoy Chakraborty
In-context learning (ICL) unfolds as large language models become capable of inferring test labels conditioned on a few labeled samples without any gradient update.
1 code implementation • 5 Feb 2023 • Vasu Goel, Dhruv Sahnan, Subhabrata Dutta, Anil Bandhakavi, Tanmoy Chakraborty
We analyze more than 32 million posts from over 6. 8 million users across three popular online social networks to investigate the interrelations between hateful behavior, information dissemination, and polarised organization mediated by echo chambers.
1 code implementation • ACL 2022 • Subhabrata Dutta, Jeevesh Juneja, Dipankar Das, Tanmoy Chakraborty
Identifying argument components from unstructured texts and predicting the relationships expressed among them are two primary steps of argument mining.
2 code implementations • 3 Jan 2022 • Subhabrata Dutta, Samiya Caur, Soumen Chakrabarti, Tanmoy Chakraborty
Detecting and labeling stance in social media text is strongly motivated by hate speech detection, poll prediction, engagement forecasting, and concerted propaganda detection.
1 code implementation • NeurIPS 2021 • Subhabrata Dutta, Tanya Gautam, Soumen Chakrabarti, Tanmoy Chakraborty
The Transformer and its variants have been proven to be efficient sequence learners in many different domains.
1 code implementation • 13 Jun 2021 • Subhabrata Dutta, Shravika Mittal, Dipankar Das, Soumen Chakrabarti, Tanmoy Chakraborty
Second, there is a measurable positive correlation between the novelty of the root content (with respect to a streaming external corpus) and the relative size of the resulting cascade.
1 code implementation • 9 Oct 2020 • Sarah Masud, Subhabrata Dutta, Sakshi Makkar, Chhavi Jain, Vikram Goyal, Amitava Das, Tanmoy Chakraborty
Meanwhile, to predict the retweet dynamics on Twitter, we propose RETINA, a novel neural architecture that incorporates exogenous influence using scaled dot-product attention.
no code implementations • 10 Aug 2019 • Subhabrata Dutta, Dipankar Das, Tanmoy Chakraborty
Unlike previous studies which model a discussion in a static manner, in the present study, we model it as a time-varying process and solve two inter-related problems -- predict which user groups will get engaged with an ongoing discussion, and forecast the growth rate of a discussion in terms of the number of comments.
no code implementations • 31 Jul 2019 • Avishek Garain, Sainik Kumar Mahata, Subhabrata Dutta
This paper presents a method to apply Natural Language Processing for normalizing numeronyms to make them understandable by humans.
no code implementations • 7 Aug 2018 • Subhabrata Dutta, Tanmoy Chakraborty, Dipankar Das
Our proposed model outperformed the previous one in terms of domain independence; without using platform-dependent structural features, our hierarchical LSTM with word relevance attention mechanism achieved F1-scores of 71\% and 66\% respectively to predict discourse roles of comments in Reddit and Facebook discussions.