Search Results for author: Megh Thakkar

Found 16 papers, 10 papers with code

DMix: Adaptive Distance-aware Interpolative Mixup

1 code implementation ACL 2022 Ramit Sawhney, Megh Thakkar, Shrey Pandit, Ritesh Soun, Di Jin, Diyi Yang, Lucie Flek

Interpolation-based regularisation methods such as Mixup, which generate virtual training samples, have proven to be effective for various tasks and modalities. We extend Mixup and propose DMix, an adaptive distance-aware interpolative Mixup that selects samples based on their diversity in the embedding space.

Data Augmentation Diversity +2

HypMix: Hyperbolic Interpolative Data Augmentation

1 code implementation EMNLP 2021 Ramit Sawhney, Megh Thakkar, Shivam Agarwal, Di Jin, Diyi Yang, Lucie Flek

Interpolation-based regularisation methods for data augmentation have proven to be effective for various tasks and modalities.

Adversarial Robustness Data Augmentation

CIAug: Equipping Interpolative Augmentation with Curriculum Learning

1 code implementation NAACL 2022 Ramit Sawhney, Ritesh Soun, Shrey Pandit, Megh Thakkar, Sarvagya Malaviya, Yuval Pinter

CIAug achieves state-of-the-art results over existing interpolative augmentation methods on 10 benchmark datasets across 4 languages in text classification and named-entity recognition tasks.

Data Augmentation named-entity-recognition +5

Too Big to Fool: Resisting Deception in Language Models

no code implementations13 Dec 2024 Mohammad Reza Samsami, Mats Leon Richter, Juan Rodriguez, Megh Thakkar, Sarath Chandar, Maxime Gasse

Large language models must balance their weight-encoded knowledge with in-context information from prompts to generate accurate responses.

Memorization

The BrowserGym Ecosystem for Web Agent Research

2 code implementations6 Dec 2024 Thibault Le Sellier De Chezelles, Maxime Gasse, Alexandre Drouin, Massimo Caccia, Léo Boisvert, Megh Thakkar, Tom Marty, Rim Assouel, Sahar Omidi Shayegan, Lawrence Keunho Jang, Xing Han Lù, Ori Yoran, Dehan Kong, Frank F. Xu, Siva Reddy, Quentin Cappart, Graham Neubig, Ruslan Salakhutdinov, Nicolas Chapados, Alexandre Lacoste

The BrowserGym ecosystem addresses the growing need for efficient evaluation and benchmarking of web agents, particularly those leveraging automation and Large Language Models (LLMs) for web interaction tasks.

Benchmarking

Combining Domain and Alignment Vectors to Achieve Better Knowledge-Safety Trade-offs in LLMs

no code implementations11 Nov 2024 Megh Thakkar, Yash More, Quentin Fournier, Matthew Riemer, Pin-Yu Chen, Amal Zouaq, Payel Das, Sarath Chandar

There is a growing interest in training domain-expert LLMs that excel in specific technical fields compared to their general-purpose instruction-tuned counterparts.

ChartGemma: Visual Instruction-tuning for Chart Reasoning in the Wild

1 code implementation4 Jul 2024 Ahmed Masry, Megh Thakkar, Aayush Bajaj, Aaryaman Kartha, Enamul Hoque, Shafiq Joty

However, existing methods suffer crucial drawbacks across two critical axes affecting the performance of chart representation models: they are trained on data generated from underlying data tables of the charts, ignoring the visual trends and patterns in chart images, and use weakly aligned vision-language backbone models for domain-specific training, limiting their generalizability when encountering charts in the wild.

Chart Understanding Decision Making +2

A Deep Dive into the Trade-Offs of Parameter-Efficient Preference Alignment Techniques

no code implementations7 Jun 2024 Megh Thakkar, Quentin Fournier, Matthew D Riemer, Pin-Yu Chen, Amal Zouaq, Payel Das, Sarath Chandar

Large language models are first pre-trained on trillions of tokens and then instruction-tuned or aligned to specific preferences.

Self-Influence Guided Data Reweighting for Language Model Pre-training

no code implementations2 Nov 2023 Megh Thakkar, Tolga Bolukbasi, Sriram Ganapathy, Shikhar Vashishth, Sarath Chandar, Partha Talukdar

Once the pre-training corpus has been assembled, all data samples in the corpus are treated with equal importance during LM pre-training.

Language Modeling Language Modelling

Towards Robust Low-Resource Fine-Tuning with Multi-View Compressed Representations

1 code implementation16 Nov 2022 Linlin Liu, Xingxuan Li, Megh Thakkar, Xin Li, Shafiq Joty, Luo Si, Lidong Bing

Due to the huge amount of parameters, fine-tuning of pretrained language models (PLMs) is prone to overfitting in the low resource scenarios.

Chart-to-Text: A Large-Scale Benchmark for Chart Summarization

2 code implementations ACL 2022 Shankar Kantharaj, Rixie Tiffany Ko Leong, Xiang Lin, Ahmed Masry, Megh Thakkar, Enamul Hoque, Shafiq Joty

We also introduce a number of state-of-the-art neural models as baselines that utilize image captioning and data-to-text generation techniques to tackle two problem variations: one assumes the underlying data table of the chart is available while the other needs to extract data from chart images.

Data-to-Text Generation Image Captioning

Cannot find the paper you are looking for? You can Submit a new open access paper.