Search Results for author: Zhengxiao Du

Found 23 papers, 17 papers with code

P-Tuning: Prompt Tuning Can Be Comparable to Fine-tuning Across Scales and Tasks

no code implementations ACL 2022 Xiao Liu, Kaixuan Ji, Yicheng Fu, Weng Tam, Zhengxiao Du, Zhilin Yang, Jie Tang

Prompt tuning, which only tunes continuous prompts with a frozen language model, substantially reduces per-task storage and memory usage at training.

Language Modeling Language Modelling

ComplexFuncBench: Exploring Multi-Step and Constrained Function Calling under Long-Context Scenario

1 code implementation17 Jan 2025 Lucen Zhong, Zhengxiao Du, Xiaohan Zhang, Haiyi Hu, Jie Tang

However, evaluating the function calling abilities of LLMs in real-world scenarios remains under-explored due to the complexity of data collection and evaluation.

GLM-4-Voice: Towards Intelligent and Human-Like End-to-End Spoken Chatbot

1 code implementation3 Dec 2024 Aohan Zeng, Zhengxiao Du, Mingdao Liu, Kedong Wang, Shengmin Jiang, Lei Zhao, Yuxiao Dong, Jie Tang

We continue pre-training from the pre-trained text language model GLM-4-9B with a combination of unsupervised speech data, interleaved speech-text data, and supervised speech-text data, scaling up to 1 trillion tokens, achieving state-of-the-art performance in both speech language modeling and spoken question answering.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +6

Scaling Speech-Text Pre-training with Synthetic Interleaved Data

1 code implementation26 Nov 2024 Aohan Zeng, Zhengxiao Du, Mingdao Liu, Lei Zhang, Shengmin Jiang, Yuxiao Dong, Jie Tang

Starting from a pre-trained language model and scaling our pre-training to 1 trillion tokens (with 600B synthetic interleaved speech-text data), we achieve state-of-the-art performance in speech language modeling and spoken question answering, improving performance on spoken questions tasks from the previous SOTA of 13% (Moshi) to 31%.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +5

VisScience: An Extensive Benchmark for Evaluating K12 Educational Multi-modal Scientific Reasoning

no code implementations10 Sep 2024 Zhihuan Jiang, Zhen Yang, Jinhao Chen, Zhengxiao Du, Weihan Wang, Bin Xu, Jie Tang

To address this gap, we meticulously construct a comprehensive benchmark, named VisScience, which is utilized to assess the multi-modal scientific reasoning across the three disciplines of mathematics, physics, and chemistry.

Question Answering Visual Question Answering

MathGLM-Vision: Solving Mathematical Problems with Multi-Modal Large Language Model

no code implementations10 Sep 2024 Zhen Yang, Jinhao Chen, Zhengxiao Du, Wenmeng Yu, Weihan Wang, Wenyi Hong, Zhihuan Jiang, Bin Xu, Jie Tang

Large language models (LLMs) have demonstrated significant capabilities in mathematical reasoning, particularly with text-based mathematical problems.

Diversity Language Modeling +3

ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline

3 code implementations3 Apr 2024 Yifan Xu, Xiao Liu, Xinghan Liu, Zhenyu Hou, Yueyan Li, Xiaohan Zhang, Zihan Wang, Aohan Zeng, Zhengxiao Du, Wenyi Zhao, Jie Tang, Yuxiao Dong

Large language models (LLMs) have shown excellent mastering of human language, but still struggle in real-world applications that require mathematical problem-solving.

Math Mathematical Problem-Solving

ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback

no code implementations1 Apr 2024 Zhenyu Hou, Yilin Niu, Zhengxiao Du, Xiaohan Zhang, Xiao Liu, Aohan Zeng, Qinkai Zheng, Minlie Huang, Hongning Wang, Jie Tang, Yuxiao Dong

The work presents our practices of aligning LLMs with human preferences, offering insights into the challenges and solutions in RLHF implementations.

Understanding Emergent Abilities of Language Models from the Loss Perspective

no code implementations23 Mar 2024 Zhengxiao Du, Aohan Zeng, Yuxiao Dong, Jie Tang

Recent studies have put into question the belief that emergent abilities in language models are exclusive to large models.

SciInstruct: a Self-Reflective Instruction Annotated Dataset for Training Scientific Language Models

1 code implementation15 Jan 2024 Dan Zhang, Ziniu Hu, Sining Zhoubian, Zhengxiao Du, Kaiyu Yang, Zihan Wang, Yisong Yue, Yuxiao Dong, Jie Tang

To bridge these gaps, we introduce SciInstruct, a suite of scientific instructions for training scientific language models capable of college-level scientific reasoning.

Math Mathematical Reasoning +1

LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding

3 code implementations28 Aug 2023 Yushi Bai, Xin Lv, Jiajie Zhang, Hongchang Lyu, Jiankai Tang, Zhidian Huang, Zhengxiao Du, Xiao Liu, Aohan Zeng, Lei Hou, Yuxiao Dong, Jie Tang, Juanzi Li

In this paper, we introduce LongBench, the first bilingual, multi-task benchmark for long context understanding, enabling a more rigorous evaluation of long context understanding.

16k Code Completion +2

AgentBench: Evaluating LLMs as Agents

1 code implementation7 Aug 2023 Xiao Liu, Hao Yu, Hanchen Zhang, Yifan Xu, Xuanyu Lei, Hanyu Lai, Yu Gu, Hangliang Ding, Kaiwen Men, Kejuan Yang, Shudan Zhang, Xiang Deng, Aohan Zeng, Zhengxiao Du, Chenhui Zhang, Sheng Shen, Tianjun Zhang, Yu Su, Huan Sun, Minlie Huang, Yuxiao Dong, Jie Tang

We present AgentBench, a multi-dimensional evolving benchmark that currently consists of 8 distinct environments to assess LLM-as-Agent's reasoning and decision-making abilities in a multi-turn open-ended generation setting.

Decision Making Instruction Following

P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks

5 code implementations14 Oct 2021 Xiao Liu, Kaixuan Ji, Yicheng Fu, Weng Lam Tam, Zhengxiao Du, Zhilin Yang, Jie Tang

Prompt tuning, which only tunes continuous prompts with a frozen language model, substantially reduces per-task storage and memory usage at training.

Language Modeling Language Modelling

GPT Understands, Too

9 code implementations18 Mar 2021 Xiao Liu, Yanan Zheng, Zhengxiao Du, Ming Ding, Yujie Qian, Zhilin Yang, Jie Tang

Prompting a pretrained language model with natural language patterns has been proved effective for natural language understanding (NLU).

Knowledge Probing Language Modeling +3

GLM: General Language Model Pretraining with Autoregressive Blank Infilling

8 code implementations ACL 2022 Zhengxiao Du, Yujie Qian, Xiao Liu, Ming Ding, Jiezhong Qiu, Zhilin Yang, Jie Tang

On a wide range of tasks across NLU, conditional and unconditional generation, GLM outperforms BERT, T5, and GPT given the same model sizes and data, and achieves the best performance from a single pretrained model with 1. 25x parameters of BERT Large , demonstrating its generalizability to different downstream tasks.

Ranked #4 on Language Modelling on WikiText-103 (using extra training data)

Abstractive Text Summarization Classification +6

Policy-Gradient Training of Fair and Unbiased Ranking Functions

1 code implementation19 Nov 2019 Himank Yadav, Zhengxiao Du, Thorsten Joachims

is an abundant and attractive source of data for learning to rank, it can produce unfair ranking policies for both exogenous and endogenous reasons.

counterfactual Decision Making +2

Cognitive Knowledge Graph Reasoning for One-shot Relational Learning

1 code implementation13 Jun 2019 Zhengxiao Du, Chang Zhou, Ming Ding, Hongxia Yang, Jie Tang

Inferring new facts from existing knowledge graphs (KG) with explainable reasoning processes is a significant problem and has received much attention recently.

Knowledge Graphs Relational Reasoning +1

Sequential Scenario-Specific Meta Learner for Online Recommendation

1 code implementation2 Jun 2019 Zhengxiao Du, Xiaowei Wang, Hongxia Yang, Jingren Zhou, Jie Tang

Our approach is based on the insight that having a good generalization from a few examples relies on both a generic model initialization and an effective strategy for adapting this model to newly arising tasks.

Few-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.