1 code implementation • • Guangtao Zeng, Wenmian Yang, Zeqian Ju, Yue Yang, Sicheng Wang, Ruisi Zhang, Meng Zhou, Jiaqi Zeng, Xiangyu Dong, Ruoyu Zhang, Hongchao Fang, Penghui Zhu, Shu Chen, Pengtao Xie
We also study the transferability of models trained on MedDialog to low-resource medical dialogue generation tasks.
Training a good deep learning model requires substantial data and computing resources, which makes the resulting neural model a valuable intellectual property.
1 code implementation • • Meng Zhou, Zechen Li, Bowen Tan, Guangtao Zeng, Wenmian Yang, Xuehai He, Zeqian Ju, Subrato Chakravorty, Shu Chen, Xingyi Yang, Yichen Zhang, Qingyang Wu, Zhou Yu, Kun Xu, Eric Xing, Pengtao Xie
Training complex dialog generation models on small datasets bears high risk of overfitting.
On these two datasets, we train several dialogue generation models based on Transformer, GPT, and BERT-GPT.