1 code implementation • 17 May 2019 • Jose Camacho-Collados, Yerai Doval, Eugenio Martínez-Cámara, Luis Espinosa-Anke, Francesco Barbieri, Steven Schockaert
Cross-lingual embeddings represent the meaning of words from different languages in the same vector space.
1 code implementation • LREC 2020 • Elmurod Kuriyozov, Yerai Doval, Carlos Gómez-Rodríguez
Our experiments confirm that the obtained bilingual dictionaries outperform previously-available ones, and that word embeddings from a low-resource language can benefit from resource-rich closely-related languages when they are aligned together.
2 code implementations • 25 Nov 2019 • Yerai Doval, Jesús Vilares, Carlos Gómez-Rodríguez
Research on word embeddings has mainly focused on improving their performance on standard corpora, disregarding the difficulties posed by noisy texts in the form of tweets and other types of non-standard writing from social media.
1 code implementation • EMNLP 2018 • Yerai Doval, Jose Camacho-Collados, Luis Espinosa-Anke, Steven Schockaert
Cross-lingual word embeddings are becoming increasingly important in multilingual NLP.
no code implementations • 3 Dec 2018 • Yerai Doval, Carlos Gómez-Rodríguez
The resulting system analyzes the text input with no word boundaries one token at a time, which can be a character or a byte, and uses the information gathered by the language model to determine if a boundary must be placed in the current position or not.
no code implementations • LREC 2020 • Yerai Doval, Jose Camacho-Collados, Luis Espinosa-Anke, Steven Schockaert
Cross-lingual word embeddings are vector representations of words in different languages where words with similar meaning are represented by similar vectors, regardless of the language.
no code implementations • 16 Oct 2019 • Yerai Doval, Jose Camacho-Collados, Luis Espinosa-Anke, Steven Schockaert
While monolingual word embeddings encode information about words in the context of a particular language, cross-lingual embeddings define a multilingual space where word embeddings from two or more languages are integrated together.
Cross-Lingual Natural Language Inference Cross-Lingual Word Embeddings +3
no code implementations • 4 Feb 2024 • Yerai Doval, Manuel Vilares, Jesús Vilares
As an answer, microtext normalization consists in transforming those non-standard microtexts into standard well-written texts as a preprocessing step, allowing traditional approaches to continue with their usual processing.