Search Results for author: Ali Modarressi

Found 10 papers, 6 papers with code

MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory

no code implementations17 Apr 2024 Ali Modarressi, Abdullatif Köksal, Ayyoob Imani, Mohsen Fayyaz, Hinrich Schütze

While current large language models (LLMs) demonstrate some capabilities in knowledge-intensive tasks, they are limited by relying on their parameters as an implicit storage mechanism.

Hallucination Language Modelling +2

DecompX: Explaining Transformers Decisions by Propagating Token Decomposition

1 code implementation5 Jun 2023 Ali Modarressi, Mohsen Fayyaz, Ehsan Aghazadeh, Yadollah Yaghoobzadeh, Mohammad Taher Pilehvar

An emerging solution for explaining Transformer-based models is to use vector-based analysis on how the representations are formed.

RET-LLM: Towards a General Read-Write Memory for Large Language Models

1 code implementation23 May 2023 Ali Modarressi, Ayyoob Imani, Mohsen Fayyaz, Hinrich Schütze

Large language models (LLMs) have significantly advanced the field of natural language processing (NLP) through their extensive parameters and comprehensive data utilization.

Question Answering

AdapLeR: Speeding up Inference by Adaptive Length Reduction

1 code implementation ACL 2022 Ali Modarressi, Hosein Mohebbi, Mohammad Taher Pilehvar

To determine the importance of each token representation, we train a Contribution Predictor for each layer using a gradient-based saliency method.

Not All Models Localize Linguistic Knowledge in the Same Place: A Layer-wise Probing on BERToids' Representations

no code implementations13 Sep 2021 Mohsen Fayyaz, Ehsan Aghazadeh, Ali Modarressi, Hosein Mohebbi, Mohammad Taher Pilehvar

Most of the recent works on probing representations have focused on BERT, with the presumption that the findings might be similar to the other models.

Cannot find the paper you are looking for? You can Submit a new open access paper.