Word Embeddings

1108 papers with code • 0 benchmarks • 52 datasets

Word embedding is the collective name for a set of language modeling and feature learning techniques in natural language processing (NLP) where words or phrases from the vocabulary are mapped to vectors of real numbers.

Techniques for learning word embeddings can include Word2Vec, GloVe, and other neural network-based approaches that train on an NLP task such as language modeling or document classification.

( Image credit: Dynamic Word Embedding for Evolving Semantic Discovery )

ProMap: Effective Bilingual Lexicon Induction via Language Model Prompting

4mekki4/promap 28 Oct 2023

We also demonstrate the effectiveness of ProMap in re-ranking results from other BLI methods such as with aligned static word embeddings.

0
28 Oct 2023

MLFMF: Data Sets for Machine Learning for Mathematical Formalization

ul-fmf/mlfmf-data NeurIPS 2023

The collection includes the largest Lean~4 library Mathlib, and some of the largest Agda libraries: the standard library, the library of univalent mathematics Agda-unimath, and the TypeTopology library.

7
24 Oct 2023

GARI: Graph Attention for Relative Isomorphism of Arabic Word Embeddings

asif6827/gari 19 Oct 2023

Bilingual Lexical Induction (BLI) is a core challenge in NLP, it relies on the relative isomorphism of individual embedding spaces.

1
19 Oct 2023

ChatGPT-guided Semantics for Zero-shot Learning

fhshubho/cgs-zsl 18 Oct 2023

Then, we enrich word vectors by combining the word embeddings from class names and descriptions generated by ChatGPT.

1
18 Oct 2023

$\textit{Swap and Predict}$ -- Predicting the Semantic Changes in Words across Corpora by Context Swapping

a1da4/svp-swap 16 Oct 2023

Intuitively, if the meaning of $w$ does not change between $\mathcal{C}_1$ and $\mathcal{C}_2$, we would expect the distributions of contextualised word embeddings of $w$ to remain the same before and after this random swapping process.

0
16 Oct 2023

Generative Adversarial Training for Text-to-Speech Synthesis Based on Raw Phonetic Input and Explicit Prosody Modelling

tiberiu44/TTS-Cube 14 Oct 2023

We describe an end-to-end speech synthesis system that uses generative adversarial training.

224
14 Oct 2023

Lightweight Adaptation of Neural Language Models via Subspace Embedding

amitkumarj441/cikm2023_subspaceembedding 16 Aug 2023

Traditional neural word embeddings are usually dependent on a richer diversity of vocabulary.

1
16 Aug 2023

3D-EX : A Unified Dataset of Definitions and Dictionary Examples

f-almeman/3d-ex 6 Aug 2023

Definitions are a fundamental building block in lexicography, linguistics and computational semantics.

3
06 Aug 2023

Circumventing Concept Erasure Methods For Text-to-Image Generative Models

nyu-dice-lab/circumventing-concept-erasure 3 Aug 2023

Text-to-image generative models can produce photo-realistic images for an extremely broad range of concepts, and their usage has proliferated widely among the general public.

7
03 Aug 2023

Beyond One-Hot-Encoding: Injecting Semantics to Drive Image Classifiers

s1m0n38/semantic-encodings 1 Aug 2023

Finally, we discuss how this approach can be further exploited in terms of explainability and adversarial robustness.

3
01 Aug 2023