Search Results for author: D. Cai

Found 1 papers, 1 papers with code

With Greater Text Comes Greater Necessity: Inference-Time Training Helps Long Text Generation

1 code implementation21 Jan 2024 Y. Wang, D. Ma, D. Cai

Our results show that: 1) Temp-Lora substantially enhances generation quality for long text, as indicated by a 13. 2% decrease in perplexity (PPL) on a subset of PG19, and a 29. 3% decrease in PPL along with a 113. 2% increase in BLEU score on a subset of GuoFeng, 2) Temp-Lora is compatible with and enhances most existing long text generation methods, and 3) Temp-Lora can greatly reduce computational costs by shortening the context window.

Language Modelling Text Generation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.