no code implementations • EMNLP 2020 • Weishi Wang, Steven C.H. Hoi, Shafiq Joty
While participants in a multi-party multi-turn conversation simultaneously engage in multiple conversation topics, existing response selection methods are developed mainly focusing on a two-party single-conversation scenario.
no code implementations • EMNLP 2021 • Tao Yu, Shafiq Joty
In this work, we propose a novel fine-tuning method based on co-training that aims to learn more generalized semantic equivalences as a complementary to multilingual language modeling using the unlabeled data in the target language.
1 code implementation • 30 Nov 2023 • Artemis Panagopoulou, Le Xue, Ning Yu, Junnan Li, Dongxu Li, Shafiq Joty, ran Xu, Silvio Savarese, Caiming Xiong, Juan Carlos Niebles
Vision-language pre-training and instruction tuning have demonstrated general-purpose capabilities in 2D visual reasoning tasks by aligning visual encoders with state-of-the-art large language models (LLMs).
no code implementations • 28 Nov 2023 • Hailin Chen, Fangkai Jiao, Xingxuan Li, Chengwei Qin, Mathieu Ravaut, Ruochen Zhao, Caiming Xiong, Shafiq Joty
Upon its release in late 2022, ChatGPT has brought a seismic shift in the entire landscape of AI, both in research and commerce.
no code implementations • 21 Nov 2023 • Bram Wallace, Meihua Dang, Rafael Rafailov, Linqi Zhou, Aaron Lou, Senthil Purushwalkam, Stefano Ermon, Caiming Xiong, Shafiq Joty, Nikhil Naik
Large language models (LLMs) are fine-tuned using human comparison data with Reinforcement Learning from Human Feedback (RLHF) methods to make them better aligned with users' preferences.
1 code implementation • 15 Nov 2023 • Yixin Liu, Alexander R. Fabbri, Jiawen Chen, Yilun Zhao, Simeng Han, Shafiq Joty, PengFei Liu, Dragomir Radev, Chien-Sheng Wu, Arman Cohan
Our study reveals that instruction controllable text summarization remains a challenging task for LLMs, since (1) all LLMs evaluated still make factual and other types of errors in their summaries; (2) all LLM-based evaluation methods cannot achieve a strong alignment with human annotators when judging the quality of candidate summaries; (3) different LLMs show large performance gaps in summary generation and evaluation.
no code implementations • 31 Oct 2023 • Wenting Zhao, Ye Liu, Tong Niu, Yao Wan, Philip S. Yu, Shafiq Joty, Yingbo Zhou, Semih Yavuz
Moreover, a significant gap in the current landscape is the absence of a realistic benchmark for evaluating the effectiveness of grounding LLMs on heterogeneous knowledge sources (e. g., knowledge base and text).
no code implementations • 28 Oct 2023 • Hailin Chen, Amrita Saha, Steven Hoi, Shafiq Joty
With the rise of powerful closed-sourced LLMs (ChatGPT, GPT-4), there are increasing interests in distilling the capabilies of close-sourced LLMs to smaller open-sourced LLMs.
no code implementations • 16 Oct 2023 • Mathieu Ravaut, Shafiq Joty, Aixin Sun, Nancy F. Chen
Large language models (LLMs) excel in zero-shot abstractive summarization tasks, delivering fluent and pertinent summaries.
no code implementations • 15 Oct 2023 • Chengwei Qin, Chen Chen, Shafiq Joty
Inspired by the learning paradigm of humans, we propose Dynamic Module Expansion and Adaptation (DMEA), which enables the model to dynamically determine the architecture for acquiring new knowledge based on task correlation and select the most similar previous tasks to facilitate adaptation to new tasks.
1 code implementation • 13 Oct 2023 • Hung Le, Hailin Chen, Amrita Saha, Akash Gokul, Doyen Sahoo, Shafiq Joty
We find that by naturally encouraging the LLM to reuse the previously developed and verified sub-modules, CodeChain can significantly boost both modularity as well as correctness of the generated solutions, achieving relative pass@1 improvements of 35% on APPS and 76% on CodeContests.
no code implementations • 3 Oct 2023 • Iva Bojic, Jessica Chen, Si Yuan Chang, Qi Chwen Ong, Shafiq Joty, Josip Car
Human evaluation plays a crucial role in Natural Language Processing (NLP) as it assesses the quality and relevance of developed systems, thereby facilitating their enhancement.
no code implementations • 29 Sep 2023 • Ansong Ni, Pengcheng Yin, Yilun Zhao, Martin Riddell, Troy Feng, Rui Shen, Stephen Yin, Ye Liu, Semih Yavuz, Caiming Xiong, Shafiq Joty, Yingbo Zhou, Dragomir Radev, Arman Cohan
Recently, large language models (LLMs), especially those that are pretrained on code, have demonstrated strong capabilities in generating programs from natural language inputs in a few-shot or even zero-shot manner.
no code implementations • 17 Sep 2023 • Kung-Hsiang Huang, Philippe Laban, Alexander R. Fabbri, Prafulla Kumar Choubey, Shafiq Joty, Caiming Xiong, Chien-Sheng Wu
In this paper, we propose a new task of summarizing diverse information encountered in multiple news articles encompassing the same event.
no code implementations • 12 Sep 2023 • Weishi Wang, Yue Wang, Shafiq Joty, Steven C. H. Hoi
Automatic program repair (APR) is crucial to reduce manual debugging efforts for developers and improve software reliability.
1 code implementation • 7 Sep 2023 • Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause, Senthil Purushwalkam, Tong Niu, Wojciech Kryściński, Lidiya Murakhovs'ka, Prafulla Kumar Choubey, Alex Fabbri, Ye Liu, Rui Meng, Lifu Tu, Meghana Bhat, Chien-Sheng Wu, Silvio Savarese, Yingbo Zhou, Shafiq Joty, Caiming Xiong
Most open-source LLMs, on the other hand, are limited in their ability to support longer sequence lengths, which is a key requirement for many tasks that require inference over an input context.
no code implementations • 24 Aug 2023 • Ye Liu, Semih Yavuz, Rui Meng, Meghana Moorthy, Shafiq Joty, Caiming Xiong, Yingbo Zhou
This paper aims to fill this gap by investigating different methods of combining retrieved passages with LLMs to enhance answer generation.
no code implementations • 6 Aug 2023 • Mathieu Ravaut, Hailin Chen, Ruochen Zhao, Chengwei Qin, Shafiq Joty, Nancy Chen
Prompt tuning (PT), a parameter-efficient technique that only tunes the additional prompt embeddings while keeping the backbone pre-trained language model (PLM) frozen, has shown promising results in language understanding tasks, especially in low-resource scenarios.
no code implementations • 20 Jun 2023 • Xuan-Phi Nguyen, Sharifah Mahani Aljunied, Shafiq Joty, Lidong Bing
Large language models (LLMs) are known to effectively perform tasks by simply observing few exemplars.
1 code implementation • 1 Jun 2023 • Fan Yin, Jesse Vig, Philippe Laban, Shafiq Joty, Caiming Xiong, Chien-Sheng Jason Wu
Large language models (LLMs) have shown impressive performance in following natural language instructions to solve unseen tasks.
no code implementations • 31 May 2023 • Iva Bojic, Qi Chwen Ong, Shafiq Joty, Josip Car
Non-communicable diseases (NCDs) are a leading cause of global deaths, necessitating a focus on primary prevention and lifestyle behavior change.
1 code implementation • 30 May 2023 • Philippe Laban, Jesse Vig, Wojciech Kryscinski, Shafiq Joty, Caiming Xiong, Chien-Sheng Wu
Text simplification research has mostly focused on sentence-level simplification, even though many desirable edits - such as adding relevant background information or reordering content - may require document-level context.
1 code implementation • 29 May 2023 • Md Tahmid Rahman Laskar, M Saiful Bari, Mizanur Rahman, Md Amran Hossen Bhuiyan, Shafiq Joty, Jimmy Xiangji Huang
The development of large language models (LLMs) such as ChatGPT has brought a lot of attention recently.
Ranked #8 on
Natural Language Inference
on ANLI test
1 code implementation • 24 May 2023 • Ahmed Masry, Parsa Kavehzadeh, Xuan Long Do, Enamul Hoque, Shafiq Joty
Charts are very popular for analyzing data, visualizing key insights and answering complex reasoning questions about data.
Ranked #11 on
Chart Question Answering
on ChartQA
(using extra training data)
1 code implementation • 24 May 2023 • Xingxuan Li, Liying Cheng, Qingyu Tan, Hwee Tou Ng, Shafiq Joty, Lidong Bing
Our preliminary experiments show that generating intermediate reasoning steps does not always boost the performance of complex temporal question-answering tasks.
1 code implementation • 23 May 2023 • Fangkai Jiao, Zhiyang Teng, Shafiq Joty, Bosheng Ding, Aixin Sun, Zhengyuan Liu, Nancy F. Chen
Existing efforts to improve logical reasoning ability of language models have predominantly relied on supervised fine-tuning, hindering generalization to new domains and/or tasks.
1 code implementation • 23 May 2023 • Philippe Laban, Wojciech Kryściński, Divyansh Agarwal, Alexander R. Fabbri, Caiming Xiong, Shafiq Joty, Chien-Sheng Wu
To address this, we propose a new protocol for inconsistency detection benchmark creation and implement it in a 10-domain benchmark called SummEdits.
no code implementations • 22 May 2023 • Xingxuan Li, Ruochen Zhao, Yew Ken Chia, Bosheng Ding, Shafiq Joty, Soujanya Poria, Lidong Bing
Specifically, CoK consists of three stages: reasoning preparation, dynamic knowledge adapting, and answer consolidation.
1 code implementation • 11 May 2023 • Han Cheol Moon, Shafiq Joty, Ruochen Zhao, Megh Thakkar, Xu Chi
Large-scale pre-trained language models have shown outstanding performance in a variety of NLP tasks.
1 code implementation • 5 May 2023 • Ruochen Zhao, Xingxuan Li, Shafiq Joty, Chengwei Qin, Lidong Bing
As large language models (LLMs) have become the norm in NLP, demonstrating good performance in generation and reasoning tasks, one of its most fatal disadvantages is the lack of factual correctness.
1 code implementation • 4 May 2023 • Xuan Long Do, Bowei Zou, Shafiq Joty, Anh Tai Tran, Liangming Pan, Nancy F. Chen, Ai Ti Aw
In addition, we propose Conv-Distinct, a novel evaluation metric for CQG, to evaluate the diversity of the generated conversation from a context.
no code implementations • 3 May 2023 • Ruochen Zhao, Shafiq Joty, Yongjie Wang, Tan Wang
The emergence of large-scale pretrained language models has posed unprecedented challenges in deriving explanations of why the model has made some predictions.
no code implementations • 3 Apr 2023 • Lifu Tu, Jin Qu, Semih Yavuz, Shafiq Joty, Wenhao Liu, Caiming Xiong, Yingbo Zhou
Our results demonstrate the strong and efficient modeling ability of NLI-based classifiers and the large cross-lingual transfer improvements achieved by our aligned prompts, particularly in few-shot settings.
1 code implementation • 2 Apr 2023 • Iva Bojic, Josef Halim, Verena Suharman, Sreeja Tar, Qi Chwen Ong, Duy Phung, Mathieu Ravaut, Shafiq Joty, Josip Car
We applied the proposed framework to four biomedical datasets and showed relative improvement of up to 33%/40% for fine-tuning of retrieval/reader models on the BioASQ dataset when using back translation to enhance the original dataset quality.
no code implementations • 20 Mar 2023 • Ruochen Zhao, Hailin Chen, Weishi Wang, Fangkai Jiao, Xuan Long Do, Chengwei Qin, Bosheng Ding, Xiaobao Guo, Minzhi Li, Xingxuan Li, Shafiq Joty
As Large Language Models (LLMs) become popular, there emerged an important trend of using multimodality to augment the LLMs' generation ability, which enables LLMs to better interact with the world.
1 code implementation • 7 Mar 2023 • Yixin Liu, Alexander R. Fabbri, Yilun Zhao, PengFei Liu, Shafiq Joty, Chien-Sheng Wu, Caiming Xiong, Dragomir Radev
Interpretability and efficiency are two important considerations for the adoption of neural automatic metrics.
2 code implementations • 6 Mar 2023 • Mohammad Abdullah Matin Khan, M Saiful Bari, Xuan Long Do, Weishi Wang, Md Rizwan Parvez, Shafiq Joty
Recently, pre-trained large language models (LLMs) have shown impressive abilities in generating codes from natural language descriptions, repairing buggy codes, translating codes between languages, and retrieving relevant code segments.
no code implementations • 16 Feb 2023 • Chengwei Qin, Qian Li, Ruochen Zhao, Shafiq Joty
Despite this, PT has been shown to rely heavily on good initialization of the prompt embeddings.
no code implementations • 31 Jan 2023 • Xiang Lin, Prathyusha Jwalapuram, Shafiq Joty
Scheduled sampling is a curriculum learning strategy that gradually exposes the model to its own predictions during training to mitigate this bias.
no code implementations • 21 Dec 2022 • M Saiful Bari, Aston Zhang, Shuai Zheng, Xingjian Shi, Yi Zhu, Shafiq Joty, Mu Li
Pre-trained large language models can efficiently interpolate human-written prompts in a natural way.
no code implementations • 20 Dec 2022 • Bosheng Ding, Chengwei Qin, Linlin Liu, Yew Ken Chia, Shafiq Joty, Boyang Li, Lidong Bing
In this paper, we evaluate the performance of GPT-3 as a data annotator by comparing it with traditional data annotation methods and analyzing its output on a range of tasks.
no code implementations • 20 Dec 2022 • Xingxuan Li, Yutong Li, Shafiq Joty, Linlin Liu, Fei Huang, Lin Qiu, Lidong Bing
On the basis of the findings, we recommended the application of more systematic and comprehensive psychological metrics to further evaluate and improve the safety of LLMs.
2 code implementations • 19 Dec 2022 • Mathieu Ravaut, Shafiq Joty, Nancy Chen
With the rise of task-specific pre-training objectives, abstractive summarization models like PEGASUS offer appealing zero-shot performance on downstream summarization tasks.
2 code implementations • 15 Dec 2022 • Yixin Liu, Alexander R. Fabbri, PengFei Liu, Yilun Zhao, Linyong Nan, Ruilin Han, Simeng Han, Shafiq Joty, Chien-Sheng Wu, Caiming Xiong, Dragomir Radev
Human evaluation is the foundation upon which the evaluation of both summarization systems and automatic metrics rests.
1 code implementation • 30 Nov 2022 • Hailin Chen, Amrita Saha, Shafiq Joty, Steven C. H. Hoi
Machine learning models usually assume i. i. d data during training and testing, but data and tasks in real world often change over time.
1 code implementation • 29 Nov 2022 • Guangsen Wang, Shafiq Joty, Junnan Li, Steven Hoi
BotSIM adopts a layered design comprising the infrastructure layer, the adaptor layer and the application layer.
1 code implementation • 22 Nov 2022 • Guangsen Wang, Samson Tan, Shafiq Joty, Gang Wu, Jimmy Au, Steven Hoi
We have open-sourced the toolkit at https://github. com/salesforce/botsim
1 code implementation • 16 Nov 2022 • Linlin Liu, Xingxuan Li, Megh Thakkar, Xin Li, Shafiq Joty, Luo Si, Lidong Bing
Due to the huge amount of parameters, fine-tuning of pretrained language models (PLMs) is prone to overfitting in the low resource scenarios.
1 code implementation • 8 Nov 2022 • Qian Li, Shafiq Joty, Daling Wang, Shi Feng, Yifei Zhang
Sparsity of formal knowledge and roughness of non-ontological construction make sparsity problem particularly prominent in Open Knowledge Graphs (OpenKGs).
1 code implementation • 17 Oct 2022 • Mathieu Ravaut, Shafiq Joty, Nancy F. Chen
To bypass this limitation, we propose a new paradigm in second-stage abstractive summarization called SummaFusion that fuses several summary candidates to produce a novel abstractive second-stage summary.
1 code implementation • 12 Oct 2022 • Shankar Kantharaj, Xuan Long Do, Rixie Tiffany Ko Leong, Jia Qing Tan, Enamul Hoque, Shafiq Joty
In the first setting, a chart and the accompanying article is provided as input to the model.
no code implementations • 23 Sep 2022 • Lingzhi Wang, Shafiq Joty, Wei Gao, Xingshan Zeng, Kam-Fai Wong
In addition to conducting experiments on a popular dataset (ReDial), we also include a multi-domain dataset (OpenDialKG) to show the effectiveness of our model.
1 code implementation • COLING 2022 • Xuan Long Do, Bowei Zou, Liangming Pan, Nancy F. Chen, Shafiq Joty, Ai Ti Aw
While previous studies mainly focus on how to model the flow and alignment of the conversation, there has been no thorough study to date on which parts of the context and history are necessary for the model.
1 code implementation • 2 Sep 2022 • Simeng Han, Hailey Schoelkopf, Yilun Zhao, Zhenting Qi, Martin Riddell, Luke Benson, Lucy Sun, Ekaterina Zubova, Yujie Qiao, Matthew Burtell, David Peng, Jonathan Fan, Yixin Liu, Brian Wong, Malcolm Sailor, Ansong Ni, Linyong Nan, Jungo Kasai, Tao Yu, Rui Zhang, Shafiq Joty, Alexander R. Fabbri, Wojciech Kryscinski, Xi Victoria Lin, Caiming Xiong, Dragomir Radev
We present FOLIO, a human-annotated, open-domain, and logically complex and diverse dataset for reasoning in natural language (NL), equipped with first order logic (FOL) annotations.
1 code implementation • 31 May 2022 • Xuan-Phi Nguyen, Shafiq Joty, Wu Kui, Ai Ti Aw
Numerous recent work on unsupervised machine translation (UMT) implies that competent unsupervised translations of low-resource and unrelated languages, such as Nepali or Sinhala, are only possible if the model is trained in a massive multilingual environment, where these low-resource languages are mixed with high-resource counterparts.
no code implementations • 25 Mar 2022 • Tasnim Mohiuddin, Philipp Koehn, Vishrav Chaudhary, James Cross, Shruti Bhosale, Shafiq Joty
In this work, we introduce a two-stage curriculum training framework for NMT where we fine-tune a base NMT model on subsets of data, selected by both deterministic scoring using pre-trained methods and online scoring that considers prediction scores of the emerging NMT model.
1 code implementation • Findings (ACL) 2022 • Ahmed Masry, Do Xuan Long, Jia Qing Tan, Shafiq Joty, Enamul Hoque
To address the unique challenges in our benchmark involving visual and logical reasoning over charts, we present two transformer-based models that combine visual features and the data table of the chart in a unified way to answer questions.
Ranked #1 on
Chart Question Answering
on RealCQA
1 code implementation • ACL 2022 • Mathieu Ravaut, Shafiq Joty, Nancy F. Chen
Sequence-to-sequence neural networks have recently achieved great success in abstractive summarization, especially through fine-tuning large pre-trained language models on the downstream dataset.
Ranked #2 on
Document Summarization
on CNN / Daily Mail
2 code implementations • ACL 2022 • Shankar Kantharaj, Rixie Tiffany Ko Leong, Xiang Lin, Ahmed Masry, Megh Thakkar, Enamul Hoque, Shafiq Joty
We also introduce a number of state-of-the-art neural models as baselines that utilize image captioning and data-to-text generation techniques to tackle two problem variations: one assumes the underlying data table of the chart is available while the other needs to extract data from chart images.
1 code implementation • ACL 2022 • Chengwei Qin, Shafiq Joty
Existing continual relation learning (CRL) methods rely on plenty of labeled training data for learning a new task, which can be hard to acquire in real scenario as getting large and representative labeled data is often expensive and time-consuming.
1 code implementation • 22 Nov 2021 • Linlin Liu, Xin Li, Ruidan He, Lidong Bing, Shafiq Joty, Luo Si
In this work, we explore methods to make better use of the multilingual annotation and language agnostic property of KG triples, and present novel knowledge based multilingual language models (KMLMs) trained directly on the knowledge triples.
1 code implementation • EMNLP 2021 • Yingzhu Zhao, Chongjia Ni, Cheung-Chi Leung, Shafiq Joty, Eng Siong Chng, Bin Ma
For model adaptation, we use a novel gradual pruning method to adapt to target speakers without changing the model architecture, which to the best of our knowledge, has never been explored in ASR.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 15 Oct 2021 • Akhilesh Deepak Gotmare, Junnan Li, Shafiq Joty, Steven C. H. Hoi
The goal of natural language semantic code search is to retrieve a semantically relevant code snippet from a fixed set of candidates using a natural language query.
1 code implementation • ICLR 2022 • Chengwei Qin, Shafiq Joty
Existing approaches to lifelong language learning rely on plenty of labeled data for learning a new task, which is hard to obtain in most real scenarios.
1 code implementation • ACL 2022 • Bosheng Ding, Junjie Hu, Lidong Bing, Sharifah Mahani Aljunied, Shafiq Joty, Luo Si, Chunyan Miao
Much recent progress in task-oriented dialogue (ToD) systems has been driven by available annotation data across multiple domains for training.
no code implementations • ACL 2022 • Prathyusha Jwalapuram, Shafiq Joty, Xiang Lin
Given the claims of improved text generation quality across various pre-trained neural models, we consider the coherence evaluation of machine generated text to be one of the principal applications of coherence models that needs to be investigated.
no code implementations • ICLR 2022 • Xuan-Phi Nguyen, Hongyu Gong, Yun Tang, Changhan Wang, Philipp Koehn, Shafiq Joty
Modern unsupervised machine translation systems mostly train their models by generating synthetic parallel training data from large unlabeled monolingual corpora of different languages through various means, such as iterative back-translation.
5 code implementations • EMNLP 2021 • Yue Wang, Weishi Wang, Shafiq Joty, Steven C. H. Hoi
We present CodeT5, a unified pre-trained encoder-decoder Transformer model that better leverages the code semantics conveyed from the developer-assigned identifiers.
Ranked #1 on
Text-to-Code Generation
on CodeXGLUE - CONCODE
no code implementations • ACL 2021 • Linlin Liu, Bosheng Ding, Lidong Bing, Shafiq Joty, Luo Si, Chunyan Miao
With the source-language data as well as the translated data, a generation-based multilingual data augmentation method is introduced to further increase diversity by generating synthetic labeled data in multiple languages.
no code implementations • 27 Jul 2021 • Xiangxi Shi, Zhonghua Wu, Guosheng Lin, Jianfei Cai, Shafiq Joty
Therefore, in this paper, we propose a memory-based Image Manipulation Network (MIM-Net), where a set of memories learned from images is introduced to synthesize the texture information with the guidance of the textual description.
4 code implementations • NeurIPS 2021 • Junnan Li, Ramprasaath R. Selvaraju, Akhilesh Deepak Gotmare, Shafiq Joty, Caiming Xiong, Steven Hoi
Most existing methods employ a transformer-based multimodal encoder to jointly model visual tokens (region-based image features) and word tokens.
Ranked #4 on
Zero-Shot Cross-Modal Retrieval
on COCO 2014
no code implementations • ACL 2021 • Thanh-Tung Nguyen, Xuan-Phi Nguyen, Shafiq Joty, XiaoLi Li
We introduce a generic seq2seq parsing framework that casts constituency parsing problems (syntactic and discourse parsing) into a series of conditional splitting decisions.
1 code implementation • 14 Jun 2021 • Xiang Lin, Simeng Han, Shafiq Joty
Advanced large-scale neural language models have led to significant success in many language generation tasks.
no code implementations • Findings (ACL) 2021 • Tasnim Mohiuddin, M Saiful Bari, Shafiq Joty
We show that AUGVIC helps to attenuate the discrepancies between relevant and distant-domain monolingual data in traditional back-translation.
1 code implementation • NAACL 2021 • Thanh-Tung Nguyen, Xuan-Phi Nguyen, Shafiq Joty, XiaoLi Li
We introduce a novel top-down end-to-end formulation of document-level discourse parsing in the Rhetorical Structure Theory (RST) framework.
Ranked #1 on
Discourse Parsing
on RST-DT
no code implementations • ACL 2021 • Samson Tan, Shafiq Joty, Kathy Baxter, Araz Taeihagh, Gregory A. Bennett, Min-Yen Kan
Questions of fairness, robustness, and transparency are paramount to address before deploying NLP systems.
1 code implementation • NAACL 2021 • Weiwen Xu, Ai Ti Aw, Yang Ding, Kui Wu, Shafiq Joty
Neural Machine Translation (NMT) has achieved significant breakthrough in performance but is known to suffer vulnerability to input perturbations.
1 code implementation • NAACL (CALCS) 2021 • Samson Tan, Shafiq Joty
Multilingual models have demonstrated impressive cross-lingual transfer performance.
1 code implementation • COLING 2022 • Linlin Liu, Thien Hai Nguyen, Shafiq Joty, Lidong Bing, Luo Si
We operationalize our framework by first proposing a novel sense-aware cross entropy loss to model word senses explicitly.
no code implementations • 28 Jan 2021 • Amrita Saha, Shafiq Joty, Steven C. H. Hoi
Neural Module Networks (NMNs) have been quite successful in incorporating explicit reasoning as learnable modules in various question answering tasks, including the most generic form of numerical reasoning over text in Machine Reading Comprehension (MRC).
no code implementations • 1 Jan 2021 • Prathyusha Jwalapuram, Barbara Rychalska, Shafiq Joty, Dominika Basaj
Despite increasing instances of machine translation (MT) systems including extrasentential context information, the evidence for translation quality improvement is sparse, especially for discourse phenomena.
no code implementations • 1 Jan 2021 • M Saiful Bari, Tasnim Mohiuddin, Shafiq Joty
Transfer learning has yielded state-of-the-art (SoTA) results in many supervised NLP tasks.
no code implementations • NeurIPS 2020 • Jiuxiang Gu, Jason Kuen, Shafiq Joty, Jianfei Cai, Vlad Morariu, Handong Zhao, Tong Sun
Structured representations of images that model visual relationships are beneficial for many vision and vision-language applications.
no code implementations • EMNLP 2020 • Bosheng Ding, Linlin Liu, Lidong Bing, Canasai Kruengkrai, Thien Hai Nguyen, Shafiq Joty, Luo Si, Chunyan Miao
Data augmentation techniques have been widely used to improve machine learning performance as they enhance the generalization capability of models.
no code implementations • NAACL 2021 • Alexander R. Fabbri, Simeng Han, Haoyuan Li, Haoran Li, Marjan Ghazvininejad, Shafiq Joty, Dragomir Radev, Yashar Mehdad
Models pretrained with self-supervised objectives on large text corpora achieve state-of-the-art performance on English text summarization tasks.
no code implementations • EMNLP 2020 • Tao Yu, Shafiq Joty
We also introduce a joint-learning objective to better capture contextual information.
1 code implementation • EMNLP 2020 • Prathyusha Jwalapuram, Shafiq Joty, Youlin Shen
Our sentence-level model shows a 0. 5 BLEU improvement on both the WMT14 and the IWSLT13 De-En testsets, while our contextual model achieves the best results, improving from 31. 81 to 32 BLEU on WMT14 De-En testset, and from 32. 10 to 33. 13 on the IWSLT13 De-En testset, with corresponding improvements in pronoun translation.
1 code implementation • EMNLP 2020 • Yifan Gao, Chien-Sheng Wu, Jingjing Li, Shafiq Joty, Steven C. H. Hoi, Caiming Xiong, Irwin King, Michael R. Lyu
Based on the learned EDU and entailment representations, we either reply to the user our final decision "yes/no/irrelevant" of the initial question, or generate a follow-up question to inquiry more information.
no code implementations • 3 Oct 2020 • Jiahui Gao, Yi Zhou, Philip L. H. Yu, Shafiq Joty, Jiuxiang Gu
In this work, we present a novel unpaired cross-lingual method to generate image captions without relying on any caption corpus in the source or the target language.
no code implementations • ECCV 2020 • Xiangxi Shi, Xu Yang, Jiuxiang Gu, Shafiq Joty, Jianfei Cai
In this paper, we propose a novel visual encoder to explicitly distinguish viewpoint changes from semantic changes in the change captioning task.
3 code implementations • Findings (EMNLP) 2021 • Ben Krause, Akhilesh Deepak Gotmare, Bryan McCann, Nitish Shirish Keskar, Shafiq Joty, Richard Socher, Nazneen Fatema Rajani
While large-scale language models (LMs) are able to imitate the distribution of natural language well enough to generate realistic text, it is difficult to control which regions of the distribution they generate.
1 code implementation • ACL 2020 • Yifan Gao, Chien-Sheng Wu, Shafiq Joty, Caiming Xiong, Richard Socher, Irwin King, Michael Lyu, Steven C. H. Hoi
The goal of conversational machine reading is to answer user questions given a knowledge base text which may require asking clarification questions.
no code implementations • ACL 2020 • Thanh-Tung Nguyen, Xuan-Phi Nguyen, Shafiq Joty, Xiao-Li Li
We propose Differentiable Window, a new neural module and general purpose component for dynamic window selection.
no code implementations • ACL 2020 • Thanh-Tung Nguyen, Xuan-Phi Nguyen, Shafiq Joty, Xiao-Li Li
We propose a novel constituency parsing model that casts the parsing problem into a series of pointing tasks.
1 code implementation • 3 Jun 2020 • Xuan-Phi Nguyen, Shafiq Joty, Thanh-Tung Nguyen, Wu Kui, Ai Ti Aw
Recent unsupervised machine translation (UMT) systems usually employ three main principles: initialization, language modeling and iterative back-translation, though they may apply them differently.
no code implementations • CL 2020 • Tasnim Mohiuddin, Shafiq Joty
Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning.
1 code implementation • 26 May 2020 • Yifan Gao, Chien-Sheng Wu, Shafiq Joty, Caiming Xiong, Richard Socher, Irwin King, Michael R. Lyu, Steven C. H. Hoi
The goal of conversational machine reading is to answer user questions given a knowledge base text which may require asking clarification questions.
1 code implementation • ACL 2020 • Samson Tan, Shafiq Joty, Min-Yen Kan, Richard Socher
Training on only perfect Standard English corpora predisposes pre-trained neural networks to discriminate against minorities from non-standard linguistic backgrounds (e. g., African American Vernacular English, Colloquial Singapore English, etc.).
no code implementations • EACL 2021 • Tasnim Mohiuddin, Prathyusha Jwalapuram, Xiang Lin, Shafiq Joty
Although coherence modeling has come a long way in developing novel models, their evaluation on downstream applications for which they are purportedly developed has largely been neglected.
1 code implementation • EMNLP 2020 • Samson Tan, Shafiq Joty, Lav R. Varshney, Min-Yen Kan
Inflectional variation is a common feature of World Englishes such as Colloquial Singapore English and African American Vernacular English.
no code implementations • 30 Apr 2020 • Prathyusha Jwalapuram, Barbara Rychalska, Shafiq Joty, Dominika Basaj
Despite increasing instances of machine translation (MT) systems including contextual information, the evidence for translation quality improvement is sparse, especially for discourse phenomena.
no code implementations • EMNLP 2020 • Tasnim Mohiuddin, M Saiful Bari, Shafiq Joty
Most of the successful and predominant methods for bilingual lexicon induction (BLI) are mapping-based, where a linear mapping function is learned with the assumption that the word embedding spaces of different languages exhibit similar geometric structures (i. e., approximately isomorphic).
Bilingual Lexicon Induction
Cross-Lingual Word Embeddings
+1
no code implementations • ACL 2021 • M Saiful Bari, Tasnim Mohiuddin, Shafiq Joty
We propose UXLA, a novel unsupervised data augmentation framework for zero-resource transfer learning scenarios.
1 code implementation • EMNLP 2020 • Yue Wang, Shafiq Joty, Michael R. Lyu, Irwin King, Caiming Xiong, Steven C. H. Hoi
By contrast, in this work, we propose VD-BERT, a simple yet effective framework of unified vision-dialog Transformer that leverages the pretrained BERT language models for Visual Dialog tasks.
no code implementations • ICLR 2020 • Xuan-Phi Nguyen, Shafiq Joty, Steven C. H. Hoi, Richard Socher
Incorporating hierarchical structures like constituency trees has been shown to be effective for various natural language processing (NLP) tasks.
no code implementations • IJCNLP 2015 • Francisco Guzman, Shafiq Joty, Lluis Marquez, Preslav Nakov
We present a novel framework for machine translation evaluation using neural networks in a pairwise setting, where the goal is to select the better translation from a pair of hypotheses, given the reference translation.
no code implementations • WS 2014 • Shafiq Joty, Francisco Guzman, Lluis Marquez, Preslav Nakov
We present novel automatic metrics for machine translation evaluation that use discourse structure and convolution kernels to compare the discourse tree of an automatic translation with that of the human reference.
1 code implementation • 22 Nov 2019 • M Saiful Bari, Shafiq Joty, Prathyusha Jwalapuram
Recently, neural methods have achieved state-of-the-art (SOTA) results in Named Entity Recognition (NER) tasks for many languages without the need for manually crafted features.
no code implementations • EMNLP 2015 • Shafiq Joty, Alberto Barrón-Cedeño, Giovanni Da San Martino, Simone Filice, Lluís Màrquez, Alessandro Moschitti, Preslav Nakov
Community question answering, a recent evolution of question answering in the Web context, allows a user to quickly consult the opinion of a number of people on a particular topic, thus taking advantage of the wisdom of the crowd.
no code implementations • 8 Nov 2019 • Simeng Han, Xiang Lin, Shafiq Joty
The resulting attention module offers an architecturally simple and empirically effective method to improve the coverage of neural text generation.
2 code implementations • NeurIPS 2020 • Xuan-Phi Nguyen, Shafiq Joty, Wu Kui, Ai Ti Aw
Our method achieves state-of-the-art BLEU scores of 30. 7 and 43. 7 in the WMT'14 English-German and English-French translation tasks, respectively.
Ranked #9 on
Machine Translation
on WMT2014 English-German
no code implementations • 25 Sep 2019 • Xuan-Phi Nguyen, Shafiq Joty, Thanh-Tung Nguyen
The attention mechanism is an indispensable component of any state-of-the-art neural machine translation system.
2 code implementations • IJCNLP 2019 • Swaraj Khadanga, Karan Aggarwal, Shafiq Joty, Jaideep Srivastava
Monitoring patients in ICU is a challenging and high-cost task.
no code implementations • IJCNLP 2019 • Han Cheol Moon, Tasnim Mohiuddin, Shafiq Joty, Xu Chi
In this paper, we propose a unified coherence model that incorporates sentence grammar, inter-sentence coherence relations, and global coherence patterns into a common neural framework.
2 code implementations • IJCNLP 2019 • Prathyusha Jwalapuram, Shafiq Joty, Irina Temnikova, Preslav Nakov
The ongoing neural revolution in machine translation has made it easier to model larger contexts beyond the sentence-level, which can potentially help resolve some discourse-level ambiguities such as pronominal anaphora, thus enabling better translations.
1 code implementation • IJCNLP 2019 • Linlin Liu, Xiang Lin, Shafiq Joty, Simeng Han, Lidong Bing
Transition-based top-down parsing with pointer networks has achieved state-of-the-art results in multiple parsing tasks, while having a linear time complexity.
no code implementations • 21 Jul 2019 • Xiangxi Shi, Jianfei Cai, Shafiq Joty, Jiuxiang Gu
With the rapid growth of video data and the increasing demands of various applications such as intelligent video search and assistance toward visually-impaired people, video captioning task has received a lot of attention recently in computer vision and natural language processing fields.
no code implementations • ACL 2019 • Jing Ma, Wei Gao, Shafiq Joty, Kam-Fai Wong
Claim verification is generally a task of verifying the veracity of a given claim, which is critical to many downstream applications.
no code implementations • ACL 2019 • Shafiq Joty, Giuseppe Carenini, Raymond Ng, Gabriel Murray
Discourse processing is a suite of Natural Language Processing (NLP) tasks to uncover linguistic structures from texts at several levels, which can support many downstream applications.
2 code implementations • ACL 2019 • Xiang Lin, Shafiq Joty, Prathyusha Jwalapuram, M Saiful Bari
We propose an efficient neural framework for sentence-level discourse analysis in accordance with Rhetorical Structure Theory (RST).
1 code implementation • NAACL 2019 • Tasnim Mohiuddin, Shafiq Joty
Adversarial training has shown impressive success in learning bilingual dictionary without any parallel data by mapping monolingual embeddings to a shared space.
no code implementations • NAACL 2019 • Tasnim Mohiuddin, Thanh-Tung Nguyen, Shafiq Joty
We address the problem of speech act recognition (SAR) in asynchronous conversations (forums, emails).
no code implementations • ICCV 2019 • Jiuxiang Gu, Shafiq Joty, Jianfei Cai, Handong Zhao, Xu Yang, Gang Wang
Most of current image captioning models heavily rely on paired image-caption datasets.
no code implementations • CL 2018 • Shafiq Joty, Tasnim Mohiuddin
Participants in an asynchronous conversation (e. g., forum, e-mail) interact with each other at different times, performing certain communicative acts, called speech acts (e. g., question, request).
no code implementations • 14 Nov 2018 • Karan Aggarwal, Shafiq Joty, Luis Fernandez-Luque, Jaideep Srivastava
Sufficient physical activity and restful sleep play a major role in the prevention and cure of many chronic conditions.
no code implementations • 28 Sep 2018 • Saravanan Thirumuruganathan, Shameem A Puthiya Parambath, Mourad Ouzzani, Nan Tang, Shafiq Joty
Entity resolution (ER) is one of the fundamental problems in data integration, where machine learning (ML) based classifiers often provide the state-of-the-art results.
no code implementations • EMNLP 2018 • Shafiq Joty, Lluis Marquez, Preslav Nakov
We address jointly two important tasks for Question Answering in community forums: given a new question, (i) find related existing questions, and (ii) find relevant answers to this new question.
no code implementations • 8 Jul 2018 • Xiangxi Shi, Jianfei Cai, Jiuxiang Gu, Shafiq Joty
In this paper, we propose a boundary-aware hierarchical language decoder for video captioning, which consists of a high-level GRU based language decoder, working as a global (caption-level) language model, and a low-level GRU based language decoder, working as a local (phrase-level) language model.
1 code implementation • ACL 2018 • Firoj Alam, Shafiq Joty, Muhammad Imran
In such scenarios, a DNN model can leverage labeled and unlabeled data from a related domain, but it has to deal with the shift in data distributions between the source and the target domains.
1 code implementation • ACL 2018 • Tasnim Mohiuddin, Shafiq Joty, Dat Tien Nguyen
We propose a novel coherence model for written asynchronous conversations (e. g., forums, emails), and show its applications in coherence assessment and thread reconstruction tasks.
no code implementations • 2 May 2018 • Firoj Alam, Shafiq Joty, Muhammad Imran
During time-critical situations such as natural disasters, rapid classification of data posted on social networks by affected people is useful for humanitarian organizations to gain situational awareness and to plan response efforts.
1 code implementation • 16 Apr 2018 • Tanay Kumar Saha, Thomas Williams, Mohammad Al Hasan, Shafiq Joty, Nicholas K. Varberg
However, existing models for learning latent representation are inadequate for obtaining the representation vectors of the vertices for different time-stamps of a dynamic network in a meaningful way.
no code implementations • ECCV 2018 • Qing Li, Qingyi Tao, Shafiq Joty, Jianfei Cai, Jiebo Luo
Most existing works in visual question answering (VQA) are dedicated to improving the accuracy of predicted answers, while disregarding the explanations.
no code implementations • ECCV 2018 • Jiuxiang Gu, Shafiq Joty, Jianfei Cai, Gang Wang
Image captioning is a multimodal task involving computer vision and natural language processing, where the goal is to learn a mapping from the image to its natural language description.
no code implementations • ICLR 2018 • Ankit Sharma, Shafiq Joty, Himanshu Kharkwal, Jaideep Srivastava
We present a number of interesting baselines, some of which adapt existing node-level embedding models to the hyperedge-level, as well as sequence based language techniques which are adapted for set structured hypergraph topology.
no code implementations • 27 Dec 2017 • Karan Aggarwal, Shafiq Joty, Luis F. Luque, Jaideep Srivastava
This is a critical barrier for the use of this new source of signal for healthcare decision making.
no code implementations • CVPR 2018 • Jiuxiang Gu, Jianfei Cai, Shafiq Joty, Li Niu, Gang Wang
Textual-visual cross-modal retrieval has been a hot research topic in both computer vision and natural language processing communities.
no code implementations • CL 2017 • Shafiq Joty, Francisco Guzmán, Lluís Màrquez, Preslav Nakov
In this article, we explore the potential of using sentence-level discourse structure for machine translation evaluation.
no code implementations • 4 Oct 2017 • Giovanni Da San Martino, Salvatore Romeo, Alberto Barron-Cedeno, Shafiq Joty, Lluis Marquez, Alessandro Moschitti, Preslav Nakov
We compare a kernel-based system with a feed-forward neural network in a scenario where a large parallel corpus is available for training a machine translation system, bilingual dictionaries, and cross-language word embeddings.
3 code implementations • 2 Oct 2017 • Muhammad Ebraheem, Saravanan Thirumuruganathan, Shafiq Joty, Mourad Ouzzani, Nan Tang
word embeddings), we present a novel ER system, called DeepER, that achieves good accuracy, high efficiency, as well as ease-of-use (i. e., much less human efforts).
Databases
no code implementations • CONLL 2017 • Shafiq Joty, Preslav Nakov, Llu{\'\i}s M{\`a}rquez, Israa Jaradat
We address the problem of cross-language adaptation for question-question similarity reranking in community question answering, with the objective to port a system trained on one input language to another input language given labeled training data for the first language and only unlabeled data for the second language.
no code implementations • 24 Jul 2017 • Dat Tien Nguyen, Shafiq Joty, Basma El Amel Boussaha, Maarten de Rijke
Discussion forums are an important source of information.
1 code implementation • ACL 2017 • Dat Tien Nguyen, Shafiq Joty
We propose a local coherence model based on a convolutional neural network that operates over the entity grid representation of a text.
no code implementations • 21 Jun 2017 • Shafiq Joty, Preslav Nakov, Lluís Màrquez, Israa Jaradat
We address the problem of cross-language adaptation for question-question similarity reranking in community question answering, with the objective to port a system trained on one input language to another input language given labeled training data for the first language and only unlabeled data for the second language.
no code implementations • COLING 2016 • Enamul Hoque, Shafiq Joty, Llu{\'\i}s M{\`a}rquez, Alberto Barr{\'o}n-Cede{\~n}o, Giovanni Da San Martino, Aless Moschitti, ro, Preslav Nakov, Salvatore Romeo, Giuseppe Carenini
We present an interactive system to provide effective and efficient search capabilities in Community Question Answering (cQA) forums.
no code implementations • COLING 2016 • Nadir Durrani, Hassan Sajjad, Shafiq Joty, Ahmed Abdelali
We present a novel fusion model for domain adaptation in Statistical Machine Translation.
1 code implementation • 25 Oct 2016 • Tanay Kumar Saha, Shafiq Joty, Naeemul Hassan, Mohammad Al Hasan
Our first approach retrofits (already trained) Sen2Vec vectors with respect to the network in two different ways: (1) using the adjacency relations of a node, and (2) using a stochastic sampling method which is more flexible in sampling neighbors of a node.
no code implementations • 18 Oct 2016 • Giovanni Da San Martino, Alberto Barrón-Cedeño, Salvatore Romeo, Alessandro Moschitti, Shafiq Joty, Fahad A. Al Obaidli, Kateryna Tymoshenko, Antonio Uva
In the case of the Arabic question re-ranking task, for the first time we applied tree kernels on syntactic trees of Arabic sentences.
no code implementations • 4 Oct 2016 • Dat Tien Nguyen, Shafiq Joty, Muhammad Imran, Hassan Sajjad, Prasenjit Mitra
During natural or man-made disasters, humanitarian response organizations look for useful information to support their decision-making processes.
no code implementations • 12 Aug 2016 • Dat Tien Nguyen, Kamela Ali Al Mannai, Shafiq Joty, Hassan Sajjad, Muhammad Imran, Prasenjit Mitra
The current state-of-the-art classification methods require a significant amount of labeled data specific to a particular event for training plus a lot of feature engineering to achieve best results.
no code implementations • 24 Jul 2016 • Aarti Sathyanarayana, Shafiq Joty, Luis Fernandez-Luque, Ferda Ofli, Jaideep Srivastava, Ahmed Elmagarmid, Shahrad Taheri, Teresa Arora
In this paper we explore the use of deep learning to build sleep quality prediction models based on actigraphy data.
no code implementations • SEMEVAL 2016 • Alberto Barr{\'o}n-Cede{\~n}o, Daniele Bonadiman, Giovanni Da San Martino, Shafiq Joty, Aless Moschitti, ro, Fahad Al Obaidli, Salvatore Romeo, Kateryna Tymoshenko, Antonio Uva
Ranked #2 on
Question Answering
on SemEvalCQA
no code implementations • SEMEVAL 2015 • Massimo Nicosia, Simone Filice, Alberto Barr{\'o}n-Cede{\~n}o, Iman Saleh, Hamdy Mubarak, Wei Gao, Preslav Nakov, Giovanni Da San Martino, Aless Moschitti, ro, Kareem Darwish, Llu{\'\i}s M{\`a}rquez, Shafiq Joty, Walid Magdy