Bilingual Lexicon Induction
34 papers with code • 0 benchmarks • 0 datasets
Translate words from one language to another.
Benchmarks
These leaderboards are used to track progress in Bilingual Lexicon Induction
Latest papers
Prix-LM: Pretraining for Multilingual Knowledge Base Construction
To achieve this, it is crucial to represent multilingual knowledge in a shared/unified space.
An Analysis of Euclidean vs. Graph-Based Framing for Bilingual Lexicon Induction from Word Embedding Spaces
Alternatively, word embeddings may be understood as nodes in a weighted graph.
Combining Static Word Embeddings and Contextual Representations for Bilingual Lexicon Induction
Bilingual Lexicon Induction (BLI) aims to map words in one language to their translations in another, and is typically through learning linear projections to align monolingual word representation spaces.
Evaluating Word Embeddings with Categorical Modularity
We find moderate to strong positive correlations between categorical modularity and performance on the monolingual tasks of sentiment analysis and word similarity calculation and on the cross-lingual task of bilingual lexicon induction both to and from English.
Cross-Lingual Word Embedding Refinement by $\ell_1$ Norm Optimisation
It is therefore recommended that this strategy be adopted as a standard for CLWE methods.
Cross-Lingual Word Embedding Refinement by $\ell_{1}$ Norm Optimisation
It is therefore recommended that this strategy be adopted as a standard for CLWE methods.
Improving the Lexical Ability of Pretrained Language Models for Unsupervised Neural Machine Translation
Successful methods for unsupervised neural machine translation (UNMT) employ crosslingual pretraining via self-supervision, often in the form of a masked language modeling or a sequence generation task, which requires the model to align the lexical- and high-level representations of the two languages.
Learning Contextualised Cross-lingual Word Embeddings and Alignments for Extremely Low-Resource Languages Using Parallel Corpora
We propose a new approach for learning contextualised cross-lingual word embeddings based on a small parallel corpus (e. g. a few hundred sentence pairs).
Semi-Supervised Bilingual Lexicon Induction with Two-way Interaction
In this paper, we propose a new semi-supervised BLI framework to encourage the interaction between the supervised signal and unsupervised alignment.
Refinement of Unsupervised Cross-Lingual Word Embeddings
In this paper, we propose a self-supervised method to refine the alignment of unsupervised bilingual word embeddings.