2 code implementations • 17 Aug 2023 • Fawaz Sammani, Nikos Deligiannis
In this work, we propose Uni-NLX, a unified framework that consolidates all NLE tasks into a single and compact multi-task model using a unified training objective of text generation.
1 code implementation • 20 Jun 2022 • Fawaz Sammani, Boris Joukovsky, Nikos Deligiannis
Contrastive learning has revolutionized the field of computer vision, learning rich representations from unlabeled data, which generalize well to diverse vision tasks.
1 code implementation • CVPR 2022 • Fawaz Sammani, Tanmoy Mukherjee, Nikos Deligiannis
Current NLE models explain the decision-making process of a vision or vision-language model (a. k. a., task model), e. g., a VQA model, via a language model (a. k. a., explanation model), e. g., GPT.
1 code implementation • CVPR 2020 • Fawaz Sammani, Luke Melas-Kyriazi
Specifically, our caption-editing model consisting of two sub-modules: (1) EditNet, a language module with an adaptive copy mechanism (Copy-LSTM) and a Selective Copy Memory Attention mechanism (SCMA), and (2) DCNet, an LSTM-based denoising auto-encoder.
no code implementations • ICLR 2020 • Fawaz Sammani, Mahmoud Elsayed, Abdelsalam Hamdi
We wish to show that starting from the third epoch, freezing weights which have no informative gradient and are less likely to be changed during training, results in a very slight drop in the overall accuracy (and in sometimes better).
1 code implementation • 7 Sep 2019 • Fawaz Sammani, Mahmoud Elsayed
Attention-based neural encoder-decoder frameworks have been widely used for image captioning.