1 code implementation • ACL 2022 • Junzhuo Li, Deyi Xiong
In this paper, we study two issues of semantic parsing approaches to conversational question answering over a large-scale knowledge base: (1) The actions defined in grammar are not sufficient to handle uncertain reasoning common in real-world scenarios.
no code implementations • 1 Apr 2024 • Chen Yang, Junzhuo Li, Xinyao Niu, Xinrun Du, Songyang Gao, Haoran Zhang, Zhaoliang Chen, Xingwei Qu, Ruibin Yuan, Yizhi Li, Jiaheng Liu, Stephen W. Huang, Shawn Yue, Wenhu Chen, Jie Fu, Ge Zhang
To address the aforementioned limitations, this paper undertakes a comprehensive comparison of model capabilities at various pretraining intermediate checkpoints.
no code implementations • 7 Nov 2023 • Shaoyang Xu, Junzhuo Li, Deyi Xiong
Multilingual pretrained language models serve as repositories of multilingual factual knowledge.
1 code implementation • 31 Oct 2023 • Xinwei Wu, Junzhuo Li, Minghui Xu, Weilong Dong, Shuangzhi Wu, Chao Bian, Deyi Xiong
The ability of data memorization and regurgitation in pretrained language models, revealed in previous studies, brings the risk of data leakage.
no code implementations • 16 Dec 2022 • Junzhuo Li, Xinwei Wu, Weilong Dong, Shuangzhi Wu, Chao Bian, Deyi Xiong
Knowledge distillation (KD) has been widely used for model compression and knowledge transfer.
no code implementations • 16 Dec 2022 • Weilong Dong, Xinwei Wu, Junzhuo Li, Shuangzhi Wu, Chao Bian, Deyi Xiong
It broadcasts the global model in the server to each client and produces pseudo data for clients so that knowledge from the global model can be explored to enhance few-shot learning of each client model.