no code implementations • EMNLP (sustainlp) 2020 • Parul Awasthy, Bishwaranjan Bhattacharjee, John Kender, Radu Florian
Transfer learning is a popular technique to learn a task using less training data and fewer compute resources.
no code implementations • 16 Mar 2023 • Aashka Trivedi, Takuma Udagawa, Michele Merler, Rameswar Panda, Yousef El-Kurdi, Bishwaranjan Bhattacharjee
This paper proposes KD-NAS, the use of Neural Architecture Search (NAS) guided by the Knowledge Distillation process to find the optimal student model for distillation from a teacher, for a given natural language task.
no code implementations • 7 Jul 2022 • John R. Kender, Bishwaranjan Bhattacharjee, Parijat Dube, Brian Belgodere
Transfer learning is a deep-learning technique that ameliorates the problem of learning when human-annotated labels are expensive and limited.
no code implementations • 20 Nov 2020 • Ulrich Finkler, Michele Merler, Rameswar Panda, Mayoore S. Jaiswal, Hui Wu, Kandan Ramakrishnan, Chun-Fu Chen, Minsik Cho, David Kung, Rogerio Feris, Bishwaranjan Bhattacharjee
Neural Architecture Search (NAS) is a powerful tool to automatically design deep neural networks for many tasks, including image classification.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Masayasu Muraoka, Tetsuya Nasukawa, Bishwaranjan Bhattacharjee
We propose a new word representation method derived from visual objects in associated images to tackle the lexical entailment task.
no code implementations • 23 Jun 2020 • Rameswar Panda, Michele Merler, Mayoore Jaiswal, Hui Wu, Kandan Ramakrishnan, Ulrich Finkler, Chun-Fu Chen, Minsik Cho, David Kung, Rogerio Feris, Bishwaranjan Bhattacharjee
The typical way of conducting large scale NAS is to search for an architectural building block on a small dataset (either using a proxy set from the large dataset or a completely different small scale dataset) and then transfer the block to a larger dataset.
no code implementations • 20 Aug 2019 • Bishwaranjan Bhattacharjee, John R. Kender, Matthew Hill, Parijat Dube, Siyu Huo, Michael R. Glass, Brian Belgodere, Sharath Pankanti, Noel Codella, Patrick Watson
We use this measure, which we call "Predict To Learn" ("P2L"), in the two very different domains of images and semantic relations, where it predicts, from a set of "source" models, the one model most likely to produce effective transfer for training a given "target" model.
no code implementations • 24 Mar 2019 • Parijat Dube, Bishwaranjan Bhattacharjee, Siyu Huo, Patrick Watson, John Kender, Brian Belgodere
Transfer learning uses trained weights from a source model as the initial weightsfor the training of a target dataset.
no code implementations • 30 Jul 2018 • Parijat Dube, Bishwaranjan Bhattacharjee, Elisabeth Petit-Bois, Matthew Hill
This is currently addressed by Transfer Learning where one learns the small data set as a transfer task from a larger source dataset.
2 code implementations • 18 Sep 2017 • Bishwaranjan Bhattacharjee, Scott Boag, Chandani Doshi, Parijat Dube, Ben Herta, Vatche Ishakian, K. R. Jayaram, Rania Khalaf, Avesh Krishna, Yu Bo Li, Vinod Muthusamy, Ruchir Puri, Yufei Ren, Florian Rosenberg, Seetharami R. Seelam, Yandong Wang, Jian Ming Zhang, Li Zhang
Deep learning driven by large neural network models is overtaking traditional machine learning methods for understanding unstructured and perceptual data domains such as speech, text, and vision.
Distributed, Parallel, and Cluster Computing