1 code implementation • 6 Sep 2023 • Pengyu Cheng, Jiawen Xie, Ke Bai, Yong Dai, Nan Du
Besides, from the perspective of data efficiency, we propose a three-stage customized RM learning scheme, then empirically verify its effectiveness on both general preference datasets and our DSP set.
no code implementations • 25 Aug 2023 • Jiawen Xie, Pengyu Cheng, Xiao Liang, Yong Dai, Nan Du
Although dominant in natural language processing, transformer-based models remain challenged by the task of long-sequence processing, because the computational cost of self-attention operations in transformers swells quadratically with the input sequence length.
no code implementations • 24 Jul 2023 • Qi Su, Na Wang, Jiawen Xie, Yinan Chen, Xiaofan Zhang
Therefore, we propose a new automatic lung lobe segmentation framework, in which we urge the model to pay attention to the area around the pulmonary fissure during the training process, which is realized by a task-specific loss function.
1 code implementation • ACL 2023 • Jiawen Xie, Qi Su, Shaoting Zhang, and Xiaofan Zhang
Most Transformer based abstractive summarization systems have a severe mismatch between training and inference, i. e., exposure bias.