no code implementations • GWC 2018 • Anupam Mondal, Dipankar Das, Erik Cambria, Sivaji Bandyopadhyay
Information extraction in the medical domain is laborious and time-consuming due to the insufficient number of domain-specific lexicons and lack of involvement of domain experts such as doctors and medical practitioners.
no code implementations • NAACL (SMM4H) 2021 • Anupam Mondal, Sainik Mahata, Monalisa Dey, Dipankar Das
The steps for pre-processing tweets, feature extraction, and the development of the machine learning models, are described extensively in the documentation.
no code implementations • EACL (DravidianLangTech) 2021 • Sainik Mahata, Dipankar Das, Sivaji Bandyopadhyay
In this work, we take up a similar challenge of developing a sentiment analysis model that can work with English-Tamil code-mixed data.
no code implementations • ICON 2020 • Sainik Mahata, Dipankar Das, Sivaji Bandyopadhyay
In the current work, we present the description of the systems submitted to a machine translation shared task organized by ICON 2020: 17th International Conference on Natural Language Processing.
no code implementations • GWC 2016 • Anupam Mondal, Dipankar Das, Erik Cambria, Sivaji Bandyopadhyay
In order to overcome the lack of medical corpora, we have developed a WordNet for Medical Events (WME) for identifying medical terms and their sense related information using a seed list.
no code implementations • SMP (ICON) 2021 • Dipankar Das
In addition, a database consists of emotional utterances of several words has also been developed as a part of this work.
no code implementations • ICON 2021 • Soumayan Majumder, Dipankar Das
We build two types of model one is language dependent and other one is language independent.
no code implementations • ICON 2021 • Rudra Dhar, Dipankar Das
Furthermore, we used different ratios of manually labeled data and weakly labeled data to train our various machine learning models.
no code implementations • 16 Jul 2023 • Dipankar Das
The current paper suggests a model and discusses how competition and collusion arise in the digital marketplace through assortment planning or assortment optimization algorithm.
no code implementations • 20 Feb 2023 • Anjan Ray Chaudhury, Dipankar Das, Sreemanta Sarkar
Decision to participate in education depends on the circumstances individual inherits and on the returns to education she expects as well.
no code implementations • 3 Oct 2022 • Dipankar Das
In this article, the researcher aims at formally deriving a rationality pattern function and thereby, the degree of rationality of the decision-maker or the reviewer in the sequential choice problem in the e-commerce markets.
no code implementations • 16 Jun 2022 • Prantik Guha, Rudra Dhar, Dipankar Das
In this paper we describe a system submitted to the INLG 2022 Generation Challenge (GenChal) on Quality Evaluation of the Low-Resource Synthetically Generated Code-Mixed Hinglish Text.
1 code implementation • ACL 2022 • Subhabrata Dutta, Jeevesh Juneja, Dipankar Das, Tanmoy Chakraborty
Identifying argument components from unstructured texts and predicting the relationships expressed among them are two primary steps of argument mining.
no code implementations • 30 Oct 2021 • Sourya Dipta Das, Ayan Basak, Soumil Mandal, Dipankar Das
Research on adversarial attacks are becoming widely popular in the recent years.
1 code implementation • 13 Jun 2021 • Subhabrata Dutta, Shravika Mittal, Dipankar Das, Soumen Chakrabarti, Tanmoy Chakraborty
Second, there is a measurable positive correlation between the novelty of the root content (with respect to a streaming external corpus) and the relative size of the resulting cascade.
no code implementations • SEMEVAL 2020 • Avishek Garain, Sainik Mahata, Dipankar Das
This linguistic phenomenon poses a great challenge to conventional NLP domains such as Sentiment Analysis, Machine Translation, and Text Summarization, to name a few.
1 code implementation • 20 Oct 2020 • Sainik Kumar Mahata, Dipankar Das, Sivaji Bandyopadhyay
Sentiment analysis has been an active area of research in the past two decades and recently, with the advent of social media, there has been an increasing demand for sentiment analysis on social media texts.
no code implementations • ICON 2019 • Tathagata Raha, Sainik Kumar Mahata, Dipankar Das, Sivaji Bandyopadhyay
The proposed system is a modular approach that starts by tagging individual tokens with their respective languages and then passes them to different POS taggers, designed for different languages (English and Bengali, in our case).
no code implementations • 28 Jul 2020 • Sainik Kumar Mahata, Amrita Chandra, Dipankar Das, Sivaji Bandyopadhyay
The preparation of raw parallel corpus, sentiment analysis of the sentences and the training of a Character Based Neural Machine Translation model using the same has been discussed extensively in this paper.
no code implementations • 24 Jul 2020 • Avishek Garain, Sainik Kumar Mahata, Dipankar Das
This linguistic phenomenon poses a great challenge to conventional NLP domains such as Sentiment Analysis, Machine Translation, and Text Summarization, to name a few.
no code implementations • 29 May 2020 • Prashant Kapil, Asif Ekbal, Dipankar Das
Moreover, the varieties in user-generated data and the presence of various forms of hate speech makes it very challenging to identify the degree and intention of the message.
no code implementations • 9 Nov 2019 • Sainik Kumar Mahata, Soumil Mandal, Dipankar Das, Sivaji Bandyopadhyay
The use of multilingualism in the new generation is widespread in the form of code-mixed data on social media, and therefore a robust translation system is required for catering to the monolingual users, as well as for easier comprehension by language processing models.
no code implementations • 17 Sep 2019 • Abhisek Kundu, Alex Heinecke, Dhiraj Kalamkar, Sudarshan Srinivasan, Eric C. Qin, Naveen K. Mellempudi, Dipankar Das, Kunal Banerjee, Bharat Kaul, Pradeep Dubey
We propose K-TanH, a novel, highly accurate, hardware efficient approximation of popular activation function TanH for Deep Learning.
no code implementations • 29 Aug 2019 • Sudarshan Srinivasan, Pradeep Janedula, Saurabh Dhoble, Sasikanth Avancha, Dipankar Das, Naveen Mellempudi, Bharat Daga, Martin Langhammer, Gregg Baeckler, Bharat Kaul
Low-precision is the first order knob for achieving higher Artificial Intelligence Operations (AI-TOPS).
no code implementations • 10 Aug 2019 • Subhabrata Dutta, Dipankar Das, Tanmoy Chakraborty
Unlike previous studies which model a discussion in a static manner, in the present study, we model it as a time-varying process and solve two inter-related problems -- predict which user groups will get engaged with an ongoing discussion, and forecast the growth rate of a discussion in terms of the number of comments.
no code implementations • WS 2018 • Sainik Kumar Mahata, Dipankar Das, Sivaji Bandyopadhyay
In the current work, we present a description of the system submitted to WMT 2018 News Translation Shared task.
no code implementations • WS 2019 • Sainik Kumar Mahata, Avishek Garain, Adityar Rayala, Dipankar Das, Sivaji Bandyopadhyay
In the current work, we present a description of the system submitted to WMT 2019 News Translation Shared task.
no code implementations • SEMEVAL 2019 • Prashant Kapil, Asif Ekbal, Dipankar Das
The three best models that performed best on individual sub tasks are stacking of CNN-Bi-LSTM with Attention, BiLSTM with POS information added with word features and Bi-LSTM for third task.
no code implementations • 29 May 2019 • Dhiraj Kalamkar, Dheevatsa Mudigere, Naveen Mellempudi, Dipankar Das, Kunal Banerjee, Sasikanth Avancha, Dharma Teja Vooturi, Nataraj Jammalamadaka, Jianyu Huang, Hector Yuen, Jiyan Yang, Jongsoo Park, Alexander Heinecke, Evangelos Georganas, Sudarshan Srinivasan, Abhisek Kundu, Misha Smelyanskiy, Bharat Kaul, Pradeep Dubey
In this paper, we discuss the flow of tensors and various key operations in mixed precision training, and delve into details of operations, such as the rounding modes for converting FP32 tensors to BFLOAT16.
no code implementations • 29 May 2019 • Naveen Mellempudi, Sudarshan Srinivasan, Dipankar Das, Bharat Kaul
Reduced precision computation for deep neural networks is one of the key areas addressing the widening compute gap driven by an exponential growth in model size.
no code implementations • 12 Dec 2018 • Sainik Kumar Mahata, Soumil Mandal, Dipankar Das, Sivaji Bandyopadhyay
All of the systems use English-Hindi and English-Bengali language pairs containing simple sentences as well as sentences of other complexity.
1 code implementation • ECCV 2018 • Apoorv Vyas, Nataraj Jammalamadaka, Xia Zhu, Dipankar Das, Bharat Kaul, Theodore L. Willke
In conjunction with the standard cross-entropy loss, we minimize the novel loss to train an ensemble of classifiers.
no code implementations • 7 Aug 2018 • Subhabrata Dutta, Tanmoy Chakraborty, Dipankar Das
Our proposed model outperformed the previous one in terms of domain independence; without using platform-dependent structural features, our hierarchical LSTM with word relevance attention mechanism achieved F1-scores of 71\% and 66\% respectively to predict discourse roles of comments in Reddit and Facebook discussions.
no code implementations • 18 Mar 2018 • Braja Gopal Patra, Dipankar Das, Amitava Das
This paper presents overview of the shared task on sentiment analysis of code-mixed data pairs of Hindi-English and Bengali-English collected from the different social media platform.
no code implementations • 11 Mar 2018 • Soumil Mandal, Sainik Kumar Mahata, Dipankar Das
To gather attention and encourage researchers to work on this crisis, we prepared gold standard Bengali-English code-mixed data with language and polarity tag for sentiment analysis purposes.
no code implementations • 10 Mar 2018 • Soumil Mandal, Sourya Dipta Das, Dipankar Das
Language identification of social media text still remains a challenging task due to properties like code-mixing and inconsistent phonetic transliterations.
no code implementations • ICLR 2018 • Dipankar Das, Naveen Mellempudi, Dheevatsa Mudigere, Dhiraj Kalamkar, Sasikanth Avancha, Kunal Banerjee, Srinivas Sridharan, Karthik Vaidyanathan, Bharat Kaul, Evangelos Georganas, Alexander Heinecke, Pradeep Dubey, Jesus Corbal, Nikita Shustrov, Roma Dubtsov, Evarist Fomenko, Vadim Pirogov
The state-of-the-art (SOTA) for mixed precision training is dominated by variants of low precision floating point operations, and in particular, FP16 accumulating into FP32 Micikevicius et al. (2017).
no code implementations • 24 Jan 2018 • Srinivas Sridharan, Karthikeyan Vaidyanathan, Dhiraj Kalamkar, Dipankar Das, Mikhail E. Smorkalov, Mikhail Shiryaev, Dheevatsa Mudigere, Naveen Mellempudi, Sasikanth Avancha, Bharat Kaul, Pradeep Dubey
The exponential growth in use of large deep neural networks has accelerated the need for training these deep neural networks in hours or even minutes.
no code implementations • 8 Jan 2018 • Soumil Mandal, Dipankar Das
We have also tested various models trained on code-mixed data, as well as English features and the highest accuracy of 72. 50% was obtained by a Support Vector Machine (SVM) model.
no code implementations • IJCNLP 2017 • Somnath Banerjee, Partha Pakray, Riyanka Manna, Dipankar Das, Alex Gelbukh, er
In this paper, we describe a deep learning framework for analyzing the customer feedback as part of our participation in the shared task on Customer Feedback Analysis at the 8th International Joint Conference on Natural Language Processing (IJCNLP 2017).
no code implementations • IJCNLP 2017 • S. Sarkar, ip, Dipankar Das, Partha Pakray
The main aim of this shared task is to choose the correct option for each multi-choice question.
no code implementations • IJCNLP 2017 • Monalisa Dey, Anupam Mondal, Dipankar Das
IJCNLP-17 Review Opinion Diversification (RevOpiD-2017) task has been designed for ranking the top-k reviews of a product from a set of reviews, which assists in identifying a summarized output to express the opinion of the entire review set.
no code implementations • RANLP 2017 • Apurba Paul, Dipankar Das
along with deep learning to classify the patterns as characters or non-characters in order to achieve decent accuracy.
no code implementations • WS 2017 • Sainik Mahata, Dipankar Das, B, Sivaji yopadhyay
A Statistical Machine Translation (SMT) system is always trained using large parallel corpus to produce effective translation.
1 code implementation • 20 Jul 2017 • Anirban Santara, Abhishek Naik, Balaraman Ravindran, Dipankar Das, Dheevatsa Mudigere, Sasikanth Avancha, Bharat Kaul
Generative Adversarial Imitation Learning (GAIL) is a state-of-the-art algorithm for learning policies when the expert's behavior is available as a fixed set of trajectories.
no code implementations • 15 Jul 2017 • Abhisek Kundu, Kunal Banerjee, Naveen Mellempudi, Dheevatsa Mudigere, Dipankar Das, Bharat Kaul, Pradeep Dubey
Aided by such an elegant trade-off between accuracy and compute, the 8-2 model (8-bit activations, ternary weights), enhanced by ternary residual edges, turns out to be sophisticated enough to achieve very high accuracy ($\sim 1\%$ drop from our FP-32 baseline), despite $\sim 1. 6\times$ reduction in model size, $\sim 26\times$ reduction in number of multiplications, and potentially $\sim 2\times$ power-performance gain comparing to 8-8 representation, on the state-of-the-art deep network ResNet-101 pre-trained on ImageNet dataset.
no code implementations • 5 Jul 2017 • Souvick Ghosh, Dipankar Das, Tanmoy Chakraborty
Analysis of these sentiments is one of the popular research areas of present day researchers.
no code implementations • 4 Jul 2017 • Souvick Ghosh, Satanu Ghosh, Dipankar Das
Also, the index can be applied to a sentence and seamlessly extended to a paragraph or an entire document.
no code implementations • 4 Jul 2017 • Souvick Ghosh, Satanu Ghosh, Dipankar Das
While some tasks deal with identifying the presence of sentiment in the text (Subjectivity analysis), other tasks aim at determining the polarity of the text categorizing them as positive, negative and neutral.
no code implementations • 2 May 2017 • Naveen Mellempudi, Abhisek Kundu, Dheevatsa Mudigere, Dipankar Das, Bharat Kaul, Pradeep Dubey
We address this by fine-tuning Resnet-50 with 8-bit activations and ternary weights at $N=64$, improving the Top-1 accuracy to within $4\%$ of the full precision result with $<30\%$ additional training overhead.
no code implementations • 31 Jan 2017 • Naveen Mellempudi, Abhisek Kundu, Dipankar Das, Dheevatsa Mudigere, Bharat Kaul
We propose a cluster-based quantization method to convert pre-trained full precision weights into ternary weights with minimal impact on the accuracy.
no code implementations • COLING 2016 • Braja Gopal Patra, Dipankar Das, B, Sivaji yopadhyay
Finally, we developed mood classification systems using Support Vector Machines and Feed Forward Neural Networks based on the features collected from audio, lyrics, and a combination of both.
no code implementations • 29 Jul 2016 • Satanu Ghosh, Souvick Ghosh, Dipankar Das
Our system demonstrated an overall accuracy of 75. 5% for token level language identification.
no code implementations • 29 Jul 2016 • Promita Maitra, Souvick Ghosh, Dipankar Das
We have used several word-based and style-based features to identify the dif-ferences between the known and unknown problems of one given set and label the unknown ones accordingly using a Random Forest based classifier.
no code implementations • 22 Feb 2016 • Dipankar Das, Sasikanth Avancha, Dheevatsa Mudigere, Karthikeyan Vaidynathan, Srinivas Sridharan, Dhiraj Kalamkar, Bharat Kaul, Pradeep Dubey
We design and implement a distributed multinode synchronous SGD algorithm, without altering hyper parameters, or compressing data, or altering algorithmic behavior.
no code implementations • 23 Jan 2014 • Tanmoy Chakraborty, Dipankar Das, Sivaji Bandyopadhyay
As a by-product of this experiment, we have started developing a standard lexicon in Bengali that serves as a productive Bengali linguistic thesaurus.