1 code implementation • ACL 2022 • Pedro Henrique Martins, Zita Marinho, Andre Martins
Transformers are unable to model long-term memories effectively, since the amount of computation they need to perform grows with the context length.
no code implementations • 24 Sep 2024 • Pedro Henrique Martins, Patrick Fernandes, João Alves, Nuno M. Guerreiro, Ricardo Rei, Duarte M. Alves, José Pombal, Amin Farajian, Manuel Faysse, Mateusz Klimaszewski, Pierre Colombo, Barry Haddow, José G. C. de Souza, Alexandra Birch, André F. T. Martins
The quality of open-weight LLMs has seen significant improvement, yet they remain predominantly focused on English.
no code implementations • 1 May 2023 • Patrick Fernandes, Aman Madaan, Emmy Liu, António Farinhas, Pedro Henrique Martins, Amanda Bertsch, José G. C. de Souza, Shuyan Zhou, Tongshuang Wu, Graham Neubig, André F. T. Martins
Many recent advances in natural language generation have been fueled by training large language models on internet-scale data.
1 code implementation • 24 May 2022 • Pedro Henrique Martins, Zita Marinho, André F. T. Martins
Semi-parametric models, which augment generation with retrieval, have led to impressive results in language modeling and machine translation, due to their ability to retrieve fine-grained information from a datastore of examples.
1 code implementation • SpaNLP (ACL) 2022 • Pedro Henrique Martins, Zita Marinho, André F. T. Martins
On the other hand, semi-parametric models have been shown to successfully perform domain adaptation by retrieving examples from an in-domain datastore (Khandelwal et al., 2021).
1 code implementation • 1 Sep 2021 • Pedro Henrique Martins, Zita Marinho, André F. T. Martins
Transformers are unable to model long-term memories effectively, since the amount of computation they need to perform grows with the context length.
Ranked #1 on
Dialogue Generation
on CMU-DoG
no code implementations • ACL (GEM) 2021 • Sebastian Gehrmann, Tosin Adewumi, Karmanya Aggarwal, Pawan Sasanka Ammanamanchi, Aremu Anuoluwapo, Antoine Bosselut, Khyathi Raghavi Chandu, Miruna Clinciu, Dipanjan Das, Kaustubh D. Dhole, Wanyu Du, Esin Durmus, Ondřej Dušek, Chris Emezue, Varun Gangal, Cristina Garbacea, Tatsunori Hashimoto, Yufang Hou, Yacine Jernite, Harsh Jhamtani, Yangfeng Ji, Shailza Jolly, Mihir Kale, Dhruv Kumar, Faisal Ladhak, Aman Madaan, Mounica Maddela, Khyati Mahajan, Saad Mahamood, Bodhisattwa Prasad Majumder, Pedro Henrique Martins, Angelina McMillan-Major, Simon Mille, Emiel van Miltenburg, Moin Nadeem, Shashi Narayan, Vitaly Nikolaev, Rubungo Andre Niyongabo, Salomey Osei, Ankur Parikh, Laura Perez-Beltrachini, Niranjan Ramesh Rao, Vikas Raunak, Juan Diego Rodriguez, Sashank Santhanam, João Sedoc, Thibault Sellam, Samira Shaikh, Anastasia Shimorina, Marco Antonio Sobrevilla Cabezudo, Hendrik Strobelt, Nishant Subramani, Wei Xu, Diyi Yang, Akhila Yerukola, Jiawei Zhou
We introduce GEM, a living benchmark for natural language Generation (NLG), its Evaluation, and Metrics.
Ranked #1 on
Extreme Summarization
on GEM-XSum
Abstractive Text Summarization
Cross-Lingual Abstractive Summarization
+5
1 code implementation • EMNLP 2020 • Pedro Henrique Martins, Zita Marinho, André F. T. Martins
Current state-of-the-art text generators build on powerful language models such as GPT-2, achieving impressive performance.
1 code implementation • 13 Feb 2020 • Pedro Henrique Martins, Vlad Niculae, Zita Marinho, André Martins
Visual attention mechanisms are widely used in multimodal tasks, as visual question answering (VQA).
no code implementations • ACL 2019 • Pedro Henrique Martins, Zita Marinho, André F. T. Martins
Named entity recognition (NER) and entity linking (EL) are two fundamentally related tasks, since in order to perform EL, first the mentions to entities have to be detected.
Ranked #12 on
Entity Linking
on AIDA-CoNLL
no code implementations • 9 Jul 2018 • Pedro Henrique Martins, Luís Custódio, Rodrigo Ventura
Using natural language to give instructions to robots is challenging, since natural language understanding is still largely an open problem.