no code implementations • 6 Jan 2025 • Andrés Villa, Juan León Alcázar, Motasem Alfarra, Vladimir Araujo, Alvaro Soto, Bernard Ghanem
Our approach, named EAGLE, is fully agnostic to the LLM or fusion module and works as a post-pretraining approach that improves the grounding and language alignment of the visual encoder.
1 code implementation • 15 Oct 2024 • Kushal Tatariya, Vladimir Araujo, Thomas Bauwens, Miryam de Lhoneux
Pixel-based language models have emerged as a compelling alternative to subword-based language modelling, particularly because they can represent virtually any script.
no code implementations • 16 Aug 2024 • Vladimir Araujo, Marie-Francine Moens, Tinne Tuytelaars
In this paper, we present L2R, a method that isolates the training of new PEFT modules to ensure their task specialization.
1 code implementation • 2 Jul 2024 • Pablo Messina, René Vidal, Denis Parra, Álvaro Soto, Vladimir Araujo
In the first stage, we propose a \textit{Fact Extractor} that leverages large language models (LLMs) to identify factual statements from well-curated domain-specific datasets.
no code implementations • 10 Jun 2024 • David Romero, Chenyang Lyu, Haryo Akbarianto Wibowo, Teresa Lynn, Injy Hamed, Aditya Nanda Kishore, Aishik Mandal, Alina Dragonetti, Artem Abzaliev, Atnafu Lambebo Tonja, Bontu Fufa Balcha, Chenxi Whitehouse, Christian Salamea, Dan John Velasco, David Ifeoluwa Adelani, David Le Meur, Emilio Villa-Cueva, Fajri Koto, Fauzan Farooqui, Frederico Belcavello, Ganzorig Batnasan, Gisela Vallejo, Grainne Caulfield, Guido Ivetta, Haiyue Song, Henok Biadglign Ademtew, Hernán Maina, Holy Lovenia, Israel Abebe Azime, Jan Christian Blaise Cruz, Jay Gala, Jiahui Geng, Jesus-German Ortiz-Barajas, Jinheon Baek, Jocelyn Dunstan, Laura Alonso Alemany, Kumaranage Ravindu Yasas Nagasinghe, Luciana Benotti, Luis Fernando D'Haro, Marcelo Viridiano, Marcos Estecha-Garitagoitia, Maria Camila Buitrago Cabrera, Mario Rodríguez-Cantelar, Mélanie Jouitteau, Mihail Mihaylov, Mohamed Fazli Mohamed Imam, Muhammad Farid Adilazuarda, Munkhjargal Gochoo, Munkh-Erdene Otgonbold, Naome Etori, Olivier Niyomugisha, Paula Mónica Silva, Pranjal Chitale, Raj Dabre, Rendi Chevi, Ruochen Zhang, Ryandito Diandaru, Samuel Cahyawijaya, Santiago Góngora, Soyeong Jeong, Sukannya Purkayastha, Tatsuki Kuribayashi, Teresa Clifford, Thanmay Jayakumar, Tiago Timponi Torrent, Toqeer Ehsan, Vladimir Araujo, Yova Kementchedjhieva, Zara Burzo, Zheng Wei Lim, Zheng Xin Yong, Oana Ignat, Joan Nwatu, Rada Mihalcea, Thamar Solorio, Alham Fikri Aji
Visual Question Answering (VQA) is an important task in multimodal AI, and it is often used to test the ability of vision-language models to understand and reason on knowledge present in both visual and textual data.
1 code implementation • 27 Mar 2024 • Nedjma Ousidhoum, Shamsuddeen Hassan Muhammad, Mohamed Abdalla, Idris Abdulmumin, Ibrahim Said Ahmad, Sanchit Ahuja, Alham Fikri Aji, Vladimir Araujo, Meriem Beloucif, Christine de Kock, Oumaima Hourrane, Manish Shrivastava, Thamar Solorio, Nirmal Surange, Krishnapriya Vishnubhotla, Seid Muhie Yimam, Saif M. Mohammad
We present the first shared task on Semantic Textual Relatedness (STR).
2 code implementations • 13 Feb 2024 • Nedjma Ousidhoum, Shamsuddeen Hassan Muhammad, Mohamed Abdalla, Idris Abdulmumin, Ibrahim Said Ahmad, Sanchit Ahuja, Alham Fikri Aji, Vladimir Araujo, Abinew Ali Ayele, Pavan Baswani, Meriem Beloucif, Chris Biemann, Sofia Bourhim, Christine de Kock, Genet Shanko Dekebo, Oumaima Hourrane, Gopichand Kanumolu, Lokesh Madasu, Samuel Rutunda, Manish Shrivastava, Thamar Solorio, Nirmal Surange, Hailegnaw Getaneh Tilaye, Krishnapriya Vishnubhotla, Genta Winata, Seid Muhie Yimam, Saif M. Mohammad
Exploring and quantifying semantic relatedness is central to representing language and holds significant implications across various NLP tasks.
1 code implementation • 20 Sep 2023 • Vladimir Araujo, Maria Mihaela Trusca, Rodrigo Tufiño, Marie-Francine Moens
In recent years, significant advancements in pre-trained language models have driven the creation of numerous non-English language variants, with a particular emphasis on encoder-only and decoder-only architectures.
no code implementations • 12 May 2023 • Vladimir Araujo, Alvaro Soto, Marie-Francine Moens
Existing question answering methods often assume that the input content (e. g., documents or videos) is always accessible to solve the task.
no code implementations • 3 Oct 2022 • Vladimir Araujo, Helena Balabin, Julio Hurtado, Alvaro Soto, Marie-Francine Moens
Lifelong language learning seeks to have models continuously learn multiple tasks in a sequential order without suffering from catastrophic forgetting.
1 code implementation • 4 Jul 2022 • Julio Hurtado, Alain Raymond-Saez, Vladimir Araujo, Vincenzo Lomonaco, Alvaro Soto, Davide Bacciu
This paper introduces Memory Outlier Elimination (MOE), a method for identifying and eliminating outliers in the memory buffer by choosing samples from label-homogeneous subpopulations.
2 code implementations • LREC 2022 • José Cañete, Sebastián Donoso, Felipe Bravo-Marquez, Andrés Carvallo, Vladimir Araujo
In this paper we present ALBETO and DistilBETO, which are versions of ALBERT and DistilBERT pre-trained exclusively on Spanish corpora.
1 code implementation • 18 Apr 2022 • Vladimir Araujo, Julio Hurtado, Alvaro Soto, Marie-Francine Moens
The ability to continuously learn remains elusive for deep learning models.
1 code implementation • LREC 2022 • Vladimir Araujo, Andrés Carvallo, Souvik Kundu, José Cañete, Marcelo Mendoza, Robert E. Mercer, Felipe Bravo-Marquez, Marie-Francine Moens, Alvaro Soto
Due to the success of pre-trained language models, versions of languages other than English have been released in recent years.
no code implementations • nlppower (ACL) 2022 • Cristóbal Eyzaguirre, Felipe del Río, Vladimir Araujo, Álvaro Soto
Large-scale pre-trained language models have shown remarkable results in diverse NLP applications.
no code implementations • EMNLP 2021 • Vladimir Araujo, Andrés Villa, Marcelo Mendoza, Marie-Francine Moens, Alvaro Soto
Current language models are usually trained using a self-supervised scheme, where the main focus is learning representations at the word or sentence level.
1 code implementation • NAACL (BioNLP) 2021 • Vladimir Araujo, Andrés Carvallo, Carlos Aspillaga, Camilo Thorne, Denis Parra
The success of pretrained word embeddings has motivated their use in the biomedical domain, with contextualized embeddings yielding remarkable results in several biomedical NLP tasks.
2 code implementations • 21 Jun 2021 • Andrés Villa, Juan-Manuel Perez-Rua, Vladimir Araujo, Juan Carlos Niebles, Victor Escorcia, Alvaro Soto
Recently, few-shot learning has received increasing interest.
no code implementations • 1 Jan 2021 • Cristobal Eyzaguirre, Felipe del Rio, Vladimir Araujo, Alvaro Soto
DACT-BERT adds an adaptive computation mechanism to the regular processing pipeline of BERT.
1 code implementation • 30 Jul 2020 • Andrés Villa, Vladimir Araujo, Francisca Cattan, Denis Parra
Our evaluation indicates that both the Transformer architecture and the contextual information are essential to get the best results for this item recommendation task.
no code implementations • WS 2020 • Vladimir Araujo, Andr{\'e}s Carvallo, Denis Parra
The success of pre-trained word embeddings of the BERT model has motivated its use in tasks in the biomedical domain.
no code implementations • WS 2020 • Patricio Cerda-Mardini, Vladimir Araujo, Alvaro Soto
We propose a multi-head attention mechanism as a blending layer in a neural network model that translates natural language to a high level behavioral language for indoor robot navigation.
no code implementations • 23 Apr 2020 • Vladimir Araujo, Andres Carvallo, Carlos Aspillaga, Denis Parra
We also show that we can significantly improve the robustness of the models by training them with adversarial examples.
no code implementations • LREC 2020 • Carlos Aspillaga, Andrés Carvallo, Vladimir Araujo
There has been significant progress in recent years in the field of Natural Language Processing thanks to the introduction of the Transformer architecture.
Natural Language Inference Natural Language Understanding +1