Search Results for author: Bingning Wang

Found 28 papers, 14 papers with code

LongReD: Mitigating Short-Text Degradation of Long-Context Large Language Models via Restoration Distillation

no code implementations11 Feb 2025 Zican Dong, Junyi Li, Jinhao Jiang, Mingyu Xu, Wayne Xin Zhao, Bingning Wang, WeiPeng Chen

To address these challenges, we propose Long Context Pre-training with Restoration Distillation (LongReD), a novel approach designed to mitigate short-text performance degradation through minimizing the distribution discrepancy between the extended and original models.

Virgo: A Preliminary Exploration on Reproducing o1-like MLLM

2 code implementations3 Jan 2025 Yifan Du, Zikang Liu, YiFan Li, Wayne Xin Zhao, Yuqi Huo, Bingning Wang, WeiPeng Chen, Zheng Liu, Zhongyuan Wang, Ji-Rong Wen

Moreover, it seems that such textual reasoning data can be even more effective than visual reasoning data in eliciting the slow-thinking capacities of MLLMs.

Language Modeling Language Modelling +1

KV Shifting Attention Enhances Language Modeling

1 code implementation29 Nov 2024 Mingyu Xu, Wei Cheng, Bingning Wang, WeiPeng Chen

In order to more efficiently implement the ability of the model's induction, we revisit the induction heads mechanism and proposed a KV shifting attention.

In-Context Learning Language Modeling +1

LLaSA: Large Language and Structured Data Assistant

no code implementations16 Nov 2024 Yao Xu, Shizhu He, Zeng Xiangrong, Jiabei Chen, Guang Liu, Bingning Wang, Jun Zhao, Kang Liu

Specifically, we represent various types of structured data in a unified hypergraph format, and use self-supervised learning to pretrain a hypergraph encoder, and a G-Former compressing encoded hypergraph representations with cross-attention.

Hypergraph representations Question Answering +1

Beyond Filtering: Adaptive Image-Text Quality Enhancement for MLLM Pretraining

1 code implementation21 Oct 2024 Han Huang, Yuqi Huo, Zijia Zhao, Haoyu Lu, Shu Wu, Bingning Wang, Qiang Liu, WeiPeng Chen, Liang Wang

A critical factor in training MLLMs is the quality of image-text pairs within multimodal pretraining datasets.

Exploring the Design Space of Visual Context Representation in Video MLLMs

1 code implementation17 Oct 2024 Yifan Du, Yuqi Huo, Kun Zhou, Zijia Zhao, Haoyu Lu, Han Huang, Wayne Xin Zhao, Bingning Wang, WeiPeng Chen, Ji-Rong Wen

Then, we explore the scaling effects in frame selection and token selection respectively, and fit the corresponding function curve by conducting extensive empirical experiments.

Language Modeling Language Modelling

Extracting and Transferring Abilities For Building Multi-lingual Ability-enhanced Large Language Models

no code implementations10 Oct 2024 Zhipeng Chen, Liang Song, Kun Zhou, Wayne Xin Zhao, Bingning Wang, WeiPeng Chen, Ji-Rong Wen

In the extraction stage, we firstly locate key neurons that are highly related to specific abilities, and then employ them to extract the transferable ability-specific weights.

Towards Event-oriented Long Video Understanding

1 code implementation20 Jun 2024 Yifan Du, Kun Zhou, Yuqi Huo, YiFan Li, Wayne Xin Zhao, Haoyu Lu, Zijia Zhao, Bingning Wang, WeiPeng Chen, Ji-Rong Wen

Leveraging an effective instruction synthesis method and an adaptive model architecture, VIM surpasses both state-of-the-art open-source models and GPT-4V on the Event-Bench.

Video Understanding

Unveiling the Flaws: Exploring Imperfections in Synthetic Data and Mitigation Strategies for Large Language Models

no code implementations18 Jun 2024 Jie Chen, Yupeng Zhang, Bingning Wang, Wayne Xin Zhao, Ji-Rong Wen, WeiPeng Chen

Synthetic data has been proposed as a solution to address the issue of high-quality data scarcity in the training of large language models (LLMs).

Instruction Following

Full-ECE: A Metric For Token-level Calibration on Large Language Models

no code implementations17 Jun 2024 Han Liu, Yupeng Zhang, Bingning Wang, WeiPeng Chen, Xiaolin Hu

Deep Neural Networks (DNNs) excel in various domains but face challenges in providing accurate uncertainty estimates, which are crucial for high-stakes applications.

MetaGPT: Merging Large Language Models Using Model Exclusive Task Arithmetic

no code implementations17 Jun 2024 Yuyan Zhou, Liang Song, Bingning Wang, WeiPeng Chen

The advent of large language models (LLMs) like GPT-4 has catalyzed the exploration of multi-task learning (MTL), in which a single model demonstrates proficiency across diverse tasks.

Computational Efficiency Task Arithmetic

Needle In A Video Haystack: A Scalable Synthetic Evaluator for Video MLLMs

1 code implementation13 Jun 2024 Zijia Zhao, Haoyu Lu, Yuqi Huo, Yifan Du, Tongtian Yue, Longteng Guo, Bingning Wang, WeiPeng Chen, Jing Liu

In this paper, we propose VideoNIAH (Video Needle In A Haystack), a benchmark construction framework through synthetic video generation.

Benchmarking Video Generation +1

Accurate and Reliable Predictions with Mutual-Transport Ensemble

no code implementations30 May 2024 Han Liu, Peng Cui, Bingning Wang, Jun Zhu, Xiaolin Hu

Deep Neural Networks (DNNs) have achieved remarkable success in a variety of tasks, especially when it comes to prediction accuracy.

Prediction

Base of RoPE Bounds Context Length

no code implementations23 May 2024 Xin Men, Mingyu Xu, Bingning Wang, Qingyu Zhang, Hongyu Lin, Xianpei Han, WeiPeng Chen

We revisit the role of RoPE in LLMs and propose a novel property of long-term decay, we derive that the \textit{base of RoPE bounds context length}: there is an absolute lower bound for the base value to obtain certain context length capability.

Position

Checkpoint Merging via Bayesian Optimization in LLM Pretraining

no code implementations28 Mar 2024 Deyuan Liu, Zecheng Wang, Bingning Wang, WeiPeng Chen, Chunshan Li, Zhiying Tu, Dianhui Chu, Bo Li, Dianbo Sui

The rapid proliferation of large language models (LLMs) such as GPT-4 and Gemini underscores the intense demand for resources during their training processes, posing significant challenges due to substantial computational and environmental costs.

Bayesian Optimization

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect

1 code implementation6 Mar 2024 Xin Men, Mingyu Xu, Qingyu Zhang, Bingning Wang, Hongyu Lin, Yaojie Lu, Xianpei Han, WeiPeng Chen

As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters.

Quantization

ChiQA: A Large Scale Image-based Real-World Question Answering Dataset for Multi-Modal Understanding

1 code implementation5 Aug 2022 Bingning Wang, Feiyang Lv, Ting Yao, Yiming Yuan, Jin Ma, Yu Luo, Haijin Liang

However, in most of the public visual question answering datasets such as VQA, CLEVR, the questions are human generated that specific to the given image, such as `What color are her eyes?'.

Image Retrieval Question Answering +2

ComQA:Compositional Question Answering via Hierarchical Graph Neural Networks

1 code implementation16 Jan 2021 Bingning Wang, Ting Yao, WeiPeng Chen, Jingfang Xu, Xiaochuan Wang

In compositional question answering, the systems should assemble several supporting evidence from the document to generate the final answer, which is more difficult than sentence-level or phrase-level QA.

Answer Selection Machine Reading Comprehension +2

SenSeNet: Neural Keyphrase Generation with Document Structure

no code implementations12 Dec 2020 Yichao Luo, Zhengyan Li, Bingning Wang, Xiaoyu Xing, Qi Zhang, Xuanjing Huang

Keyphrase Generation (KG) is the task of generating central topics from a given document or literary work, which captures the crucial information necessary to understand the content.

Inductive Bias Keyphrase Generation +1

Sogou Machine Reading Comprehension Toolkit

1 code implementation28 Mar 2019 Jindou Wu, Yunlun Yang, Chao Deng, Hongyi Tang, Bingning Wang, Haoze Sun, Ting Yao, Qi Zhang

In this paper, we present a Sogou Machine Reading Comprehension (SMRC) toolkit that can be used to provide the fast and efficient development of modern machine comprehension models, including both published models and original prototypes.

Machine Reading Comprehension

Cannot find the paper you are looking for? You can Submit a new open access paper.