no code implementations • 10 Apr 2024 • Yongqiang Ma, Lizhi Qing, Jiawei Liu, Yangyang Kang, Yue Zhang, Wei Lu, Xiaozhong Liu, Qikai Cheng
Therefore, our study shifts the focus from model-centered to human-centered evaluation in the context of AI-powered writing assistance applications.
1 code implementation • 16 Feb 2024 • Yinpeng Liu, Jiawei Liu, Xiang Shi, Qikai Cheng, Wei Lu
We advocate the few-shot in-context curriculum learning (ICCL), a simple but effective demonstration ordering method for ICL, which implies gradually increasing the complexity of prompt demonstrations during the inference process.
no code implementations • 19 Oct 2023 • Xiang Shi, Jiawei Liu, Yinpeng Liu, Qikai Cheng, Wei Lu
The advent of Large Language Models (LLMs) has shown the potential to improve relevance and provide direct answers in web searches.
no code implementations • 5 May 2023 • Jiawei Liu, Zi Xiong, Yi Jiang, Yongqiang Ma, Wei Lu, Yong Huang, Qikai Cheng
Inspired by recent advancement in prompt learning, in this paper, we propose the Mix Prompt Tuning (MPT), which is a semi-supervised method to alleviate the dependence on annotated data and improve the performance of multi-granularity academic function recognition tasks with a small number of labeled examples.
no code implementations • 24 Jan 2023 • Yongqiang Ma, Jiawei Liu, Fan Yi, Qikai Cheng, Yong Huang, Wei Lu, Xiaozhong Liu
We find that there exists a "writing style" gap between AI-generated scientific text and human-written scientific text.
no code implementations • 7 Sep 2022 • Xin Li, Xuli Tang, Qikai Cheng
We extracted ninety-one paper features from three dimensions as the input of the model, including twenty-one features in the paper dimension, thirty-five in the reference dimension, and thirty-five in the citing paper dimension.