Cross-Lingual Paraphrase Identification

5 papers with code • 1 benchmarks • 2 datasets

This task has no description! Would you like to contribute one?

Most implemented papers

ByT5: Towards a token-free future with pre-trained byte-to-byte models

google-research/byt5 28 May 2021

Most widely-used pre-trained language models operate on sequences of tokens corresponding to word or subword units.

Rethinking embedding coupling in pre-trained language models

PaddlePaddle/PaddleNLP ICLR 2021

We re-evaluate the standard practice of sharing weights between input and output embeddings in state-of-the-art pre-trained language models.

mGPT: Few-Shot Learners Go Multilingual

ai-forever/mgpt 15 Apr 2022

Recent studies report that autoregressive language models can successfully solve many NLP tasks via zero- and few-shot learning paradigms, which opens up new possibilities for using the pre-trained language models.

PromptBench: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts

microsoft/promptbench 7 Jun 2023

The increasing reliance on Large Language Models (LLMs) across academia and industry necessitates a comprehensive understanding of their robustness to prompts.

Do Multilingual Language Models Think Better in English?

juletx/self-translate 2 Aug 2023

In this work, we introduce a new approach called self-translate, which overcomes the need of an external translation system by leveraging the few-shot translation capabilities of multilingual language models.