no code implementations • 25 Feb 2025 • Jianhao Yan, Yun Luo, Yue Zhang
Meta-evaluation shows that the LLM-based refuter could generate more human-like refutations and the evaluators could assign scores with high correlation with humans.
no code implementations • 21 Feb 2025 • Zhilin Wang, Yafu Li, Jianhao Yan, Yu Cheng, Yue Zhang
Dynamical systems theory provides a framework for analyzing iterative processes and evolution over time.
1 code implementation • 21 Nov 2024 • Jianhao Yan, Pingchuan Yan, Yulong Chen, Jing Li, Xianchao Zhu, Yue Zhang
This study presents a comprehensive evaluation of GPT-4's translation capabilities compared to human translators of varying expertise levels.
no code implementations • 12 Oct 2024 • Jianhao Yan, Futing Wang, Yun Luo, Yafu Li, Yue Zhang
Large language models (LLMs) have revolutionized knowledge storage and retrieval, but face challenges with conflicting and outdated information.
1 code implementation • 12 Oct 2024 • Futing Wang, Jianhao Yan, Yue Zhang, Tao Lin
By externally storing and reusing vectors that represent in-context learned capabilities, \alg not only demonstrates the potential to operate modular capabilities but also significantly enhances the performance, versatility, adaptability, and scalability of large language models.
1 code implementation • 16 Aug 2024 • Yulong Chen, Yang Liu, Jianhao Yan, Xuefeng Bai, Ming Zhong, Yinghao Yang, ZiYi Yang, Chenguang Zhu, Yue Zhang
We then build a benchmark, SC-G4, consisting of 1, 835 instances generated by GPT-4 using these patterns, with human-annotated gold responses.
no code implementations • 4 Jul 2024 • Jianhao Yan, Pingchuan Yan, Yulong Chen, Judy Li, Xianchao Zhu, Yue Zhang
This study comprehensively evaluates the translation quality of Large Language Models (LLMs), specifically GPT-4, against human translators of varying expertise levels across multiple language pairs and domains.
1 code implementation • 21 May 2024 • Yafu Li, Huajian Zhang, Jianhao Yan, Yongjing Yin, Yue Zhang
Recent advances have made non-autoregressive (NAT) translation comparable to autoregressive methods (AT).
1 code implementation • 21 Feb 2024 • Jianhao Yan, Yun Luo, Yue Zhang
The application scope of large language models (LLMs) is increasingly expanding.
no code implementations • 21 Feb 2024 • Jianhao Yan, Futing Wang, Yafu Li, Yue Zhang
Large language models (LLMs) trained on vast corpora suffer from inevitable stereotype biases.
1 code implementation • 30 Oct 2023 • Chiyu Song, Zhanchao Zhou, Jianhao Yan, Yuejiao Fei, Zhenzhong Lan, Yue Zhang
Instruction tuning is a burgeoning method to elicit the general intelligence of Large Language Models (LLMs).
1 code implementation • 30 Sep 2023 • Jianhao Yan, Jin Xu, Chiyu Song, Chenming Wu, Yafu Li, Yue Zhang
This paper explores the elusive mechanism underpinning in-context learning in Large Language Models (LLMs).
1 code implementation • 8 Jul 2023 • Yulong Chen, Huajian Zhang, Yijie Zhou, Xuefeng Bai, Yueguan Wang, Ming Zhong, Jianhao Yan, Yafu Li, Judy Li, Michael Zhu, Yue Zhang
Additionally, based on the same intuition, we propose a 2-Step method, which takes both conversation and summary as input to simulate human annotation process.
1 code implementation • 20 Jun 2023 • Yafu Li, Leyang Cui, Jianhao Yan, Yongjing Yin, Wei Bi, Shuming Shi, Yue Zhang
Most existing text generation models follow the sequence-to-sequence paradigm.
1 code implementation • 22 May 2023 • Guangsheng Bao, Zhiyang Teng, Hao Zhou, Jianhao Yan, Yue Zhang
However, current NAT models still have a significant performance gap compared to their AT counterparts.
no code implementations • 8 Dec 2022 • Jianhao Yan, Jin Xu, Fandong Meng, Jie zhou, Yue Zhang
In this work, we show that the issue arises from the un-consistency of label smoothing on the token-level and sequence-level distributions.
no code implementations • 25 Jun 2022 • Jianhao Yan, Fandong Meng, Jie zhou
Hallucination, one kind of pathological translations that bothers Neural Machine Translation, has recently drawn much attention.
2 code implementations • 6 Jun 2022 • Jin Xu, Xiaojiang Liu, Jianhao Yan, Deng Cai, Huayang Li, Jian Li
While large-scale neural language models, such as GPT2 and BART, have achieved impressive results on various text generation tasks, they tend to get stuck in undesirable sentence-level loops with maximization-based decoding algorithms (\textit{e. g.}, greedy search).
no code implementations • 29 Jun 2021 • Jianhao Yan, Chenming Wu, Fandong Meng, Jie zhou
Current evaluation of an NMT system is usually built upon a heuristic decoding algorithm (e. g., beam search) and an evaluation metric assessing similarity between the translation and golden reference.
1 code implementation • ACL 2021 • Fusheng Wang, Jianhao Yan, Fandong Meng, Jie zhou
As an active research field in NMT, knowledge distillation is widely applied to enhance the model's performance by transferring teacher model's knowledge on each training sample.
1 code implementation • EMNLP 2020 • Jianhao Yan, Fandong Meng, Jie zhou
Transformer models achieve remarkable success in Neural Machine Translation.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Lin Qiao, Jianhao Yan, Fandong Meng, Zhendong Yang, Jie zhou
Therefore, we propose a novel Sentiment-Controllable topic-to-essay generator with a Topic Knowledge Graph enhanced decoder, named SCTKG, which is based on the conditional variational autoencoder (CVAE) framework.
no code implementations • WMT (EMNLP) 2020 • Fandong Meng, Jianhao Yan, Yijin Liu, Yuan Gao, Xianfeng Zeng, Qinsong Zeng, Peng Li, Ming Chen, Jie zhou, Sifan Liu, Hao Zhou
We participate in the WMT 2020 shared news translation task on Chinese to English.
no code implementations • 15 Jul 2020 • Jianhao Yan, Fandong Meng, Jie zhou
Though remarkable successes have been achieved by Neural Machine Translation (NMT) in recent years, it still suffers from the inadequate-translation problem.
no code implementations • 21 Apr 2020 • Canxiang Yan, Jianhao Yan, Yangyin Xu, Cheng Niu, Jie zhou
Static knowledge graph has been incorporated extensively into sequence-to-sequence framework for text generation.
1 code implementation • NAACL 2019 • Jianhao Yan, Lin He, Ruqin Huang, Jian Li, Ying Liu
This paper formulates the problem of relation extraction with temporal reasoning and proposes a solution to predict whether two given entities participate in a relation at a given time spot.