no code implementations • 10 Apr 2024 • Chenyang An, Zhibo Chen, Qihao Ye, Emily First, Letian Peng, Jiayun Zhang, Zihan Wang, Sorin Lerner, Jingbo Shang
Recent advances in Automated Theorem Proving have shown the effectiveness of leveraging a (large) language model that generates tactics (i. e. proof steps) to search through proof states.
1 code implementation • 30 Mar 2024 • Letian Peng, Zilong Wang, Feng Yao, Zihan Wang, Jingbo Shang
We construct the distillation dataset via sampling sentences from language model pre-training datasets (e. g., OpenWebText in our implementation) and prompting an LLM to identify the typed spans of "important information".
1 code implementation • 15 Feb 2024 • Letian Peng, Yuwei Zhang, Zilong Wang, Jayanth Srinivasa, Gaowen Liu, Zihan Wang, Jingbo Shang
This work aims to build a text embedder that can capture characteristics of texts specified by user instructions.
1 code implementation • 6 Nov 2023 • Letian Peng, Zihan Wang, Jingbo Shang
We study the named entity recognition (NER) problem under the extremely weak supervision (XWS) setting, where only one example entity per type is given in a context-free way.
1 code implementation • 3 Nov 2023 • Letian Peng, Zilong Wang, Hang Liu, Zihan Wang, Jingbo Shang
With the rapid development of the internet, online social media welcomes people with different backgrounds through its diverse content.
1 code implementation • 14 Jul 2023 • Letian Peng, Yuwei Zhang, Jingbo Shang
In this paper, we propose a novel method, Chain-of-Thoughts Attribute Manipulation (CoTAM), to guide few-shot learning by carefully crafted data from Large Language Models (LLMs).
no code implementations • 23 Aug 2022 • Letian Peng, Zuchao Li, Hai Zhao
In detail, it works on PLMs according to the Replaced Token Detection (RTD) pre-training objective in ELECTRA, in which the corruption detection objective reflects the confidence on contextual integrity that is more relevant to commonsense reasoning than existing probability.
1 code implementation • 30 Apr 2022 • Letian Peng, Zuchao Li, Hai Zhao
We report the performance of DeBERTaV3 on CommonsenseQA in this report.
no code implementations • 4 Jan 2022 • Jiajia Li, Letian Peng, Ping Wang, Zuchao Li, Xueyi Li, Hai Zhao
As the model training on information from users is likely to invade personal privacy, many methods have been proposed to block the learning and memorizing of the sensitive data in raw texts.
no code implementations • 29 Oct 2021 • Letian Peng, Zuchao Li, Hai Zhao
Unsupervised constituency parsing has been explored much but is still far from being solved.
1 code implementation • 4 Oct 2021 • Letian Peng, Zuchao Li, Hai Zhao
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
no code implementations • 14 Sep 2021 • Letian Peng, Zuchao Li, Hai Zhao
Attention scorers have achieved success in parsing tasks like semantic and syntactic dependency parsing.