Search Results for author: Mahdi Namazifar

Found 19 papers, 6 papers with code

Enhancing Knowledge Selection for Grounded Dialogues via Document Semantic Graphs

no code implementations NAACL 2022 Sha Li, Mahdi Namazifar, Di Jin, Mohit Bansal, Heng Ji, Yang Liu, Dilek Hakkani-Tur

In this work, we propose to automatically convert the background knowledge documents into document semantic graphs and then perform knowledge selection over such graphs.

Multi-Task Learning Response Generation +1

Data-Efficient Alignment of Large Language Models with Human Feedback Through Natural Language

no code implementations24 Nov 2023 Di Jin, Shikib Mehri, Devamanyu Hazarika, Aishwarya Padmakumar, Sungjin Lee, Yang Liu, Mahdi Namazifar

Learning from human feedback is a prominent technique to align the output of large language models (LLMs) with human expectations.

KILM: Knowledge Injection into Encoder-Decoder Language Models

1 code implementation17 Feb 2023 Yan Xu, Mahdi Namazifar, Devamanyu Hazarika, Aishwarya Padmakumar, Yang Liu, Dilek Hakkani-Tür

Large pre-trained language models (PLMs) have been shown to retain implicit knowledge within their parameters.

Entity Disambiguation

Role of Bias Terms in Dot-Product Attention

no code implementations16 Feb 2023 Mahdi Namazifar, Devamanyu Hazarika, Dilek Hakkani-Tur

Moreover, we argue that the bias term of the value linear transformation has a more prominent role than that of the bias term of the query linear transformation.

Language Modelling Natural Language Understanding +1

Inducer-tuning: Connecting Prefix-tuning and Adapter-tuning

1 code implementation26 Oct 2022 Yifan Chen, Devamanyu Hazarika, Mahdi Namazifar, Yang Liu, Di Jin, Dilek Hakkani-Tur

Prefix-tuning, or more generally continuous prompt tuning, has become an essential paradigm of parameter-efficient transfer learning.

Language Modelling Natural Language Understanding +1

Enhanced Knowledge Selection for Grounded Dialogues via Document Semantic Graphs

no code implementations15 Jun 2022 Sha Li, Mahdi Namazifar, Di Jin, Mohit Bansal, Heng Ji, Yang Liu, Dilek Hakkani-Tur

Providing conversation models with background knowledge has been shown to make open-domain dialogues more informative and engaging.

Multi-Task Learning Response Generation +1

Zero-Shot Controlled Generation with Encoder-Decoder Transformers

no code implementations11 Jun 2021 Devamanyu Hazarika, Mahdi Namazifar, Dilek Hakkani-Tür

In this work, we propose novel approaches for controlling encoder-decoder transformer-based NLG models in zero-shot.

Document Summarization Machine Translation +1

Warped Language Models for Noise Robust Language Understanding

no code implementations3 Nov 2020 Mahdi Namazifar, Gokhan Tur, Dilek Hakkani Tür

The insertion and drop modification of the input text during training of WLM resemble the types of noise due to Automatic Speech Recognition (ASR) errors, and as a result WLMs are likely to be more robust to ASR noise.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

Joint Contextual Modeling for ASR Correction and Language Understanding

no code implementations28 Jan 2020 Yue Weng, Sai Sumanth Miryala, Chandra Khatri, Runze Wang, Huaixiu Zheng, Piero Molino, Mahdi Namazifar, Alexandros Papangelis, Hugh Williams, Franziska Bell, Gokhan Tur

As a baseline approach, we trained task-specific Statistical Language Models (SLM) and fine-tuned state-of-the-art Generalized Pre-training (GPT) Language Model to re-rank the n-best ASR hypotheses, followed by a model to identify the dialog act and slots.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

Plato Dialogue System: A Flexible Conversational AI Research Platform

4 code implementations17 Jan 2020 Alexandros Papangelis, Mahdi Namazifar, Chandra Khatri, Yi-Chia Wang, Piero Molino, Gokhan Tur

Plato has been designed to be easy to understand and debug and is agnostic to the underlying learning frameworks that train each component.

Spoken Dialogue Systems

Flexibly-Structured Model for Task-Oriented Dialogues

1 code implementation WS 2019 Lei Shu, Piero Molino, Mahdi Namazifar, Hu Xu, Bing Liu, Huaixiu Zheng, Gokhan Tur

It is based on a simple and practical yet very effective sequence-to-sequence approach, where language understanding and state tracking tasks are modeled jointly with a structured copy-augmented sequential decoder and a multi-label decoder for each slot.

Task-Oriented Dialogue Systems Text Generation

Named Entity Sequence Classification

no code implementations6 Dec 2017 Mahdi Namazifar

We frame NESC as a binary classification problem and we use NER as well as recurrent neural networks to find the probability of candidate named entity is a real named entity.

Binary Classification Classification +4

Cannot find the paper you are looking for? You can Submit a new open access paper.