Search Results for author: Rocktim Jyoti Das

Found 5 papers, 3 papers with code

Factuality of Large Language Models in the Year 2024

no code implementations4 Feb 2024 Yuxia Wang, Minghan Wang, Muhammad Arslan Manzoor, Fei Liu, Georgi Georgiev, Rocktim Jyoti Das, Preslav Nakov

Large language models (LLMs), especially when instruction-tuned for chat, have become part of our daily lives, freeing people from the process of searching, extracting, and integrating information from multiple sources by offering a straightforward answer to a variety of questions in a single place.

Text Generation

Beyond Size: How Gradients Shape Pruning Decisions in Large Language Models

2 code implementations8 Nov 2023 Rocktim Jyoti Das, MingJie Sun, Liqun Ma, Zhiqiang Shen

GBLM-Pruner leverages the first-order term of the Taylor expansion, operating in a training-free manner by harnessing properly normalized gradients from a few calibration samples to determine the pruning metric, and substantially outperforms competitive counterparts like SparseGPT and Wanda in multiple benchmarks.

Language Modelling Network Pruning

DKAF: KB Arbitration for Learning Task-Oriented Dialog Systems with Dialog-KB Inconsistencies

1 code implementation26 May 2023 Vishal Vivek Saley, Rocktim Jyoti Das, Dinesh Raghu, Mausam

In this work, we define the novel problem of learning a TOD agent with dialog-KB inconsistencies in the training data.

Exploring Distributional Shifts in Large Language Models for Code Analysis

no code implementations16 Mar 2023 Shushan Arakelyan, Rocktim Jyoti Das, Yi Mao, Xiang Ren

We systematically study how three large language models with code capabilities - CodeT5, Codex, and ChatGPT - generalize to out-of-domain data.

Code Generation Code Summarization

Cannot find the paper you are looking for? You can Submit a new open access paper.