Search Results for author: Zenan Zhou

Found 16 papers, 11 papers with code

Beyond Sight: Towards Cognitive Alignment in LVLM via Enriched Visual Knowledge

no code implementations25 Nov 2024 Yaqi Zhao, Yuanyang Yin, Lin Li, MingAn Lin, Victor Shea-Jay Huang, Siwei Chen, WeiPeng Chen, Baoqun Yin, Zenan Zhou, Wentao Zhang

Specifically, the VE's representation of visual information may not fully align with LLM's cognitive framework, leading to a mismatch where visual features exceed the language model's interpretive range.

Landmark Recognition Large Language Model

VersaTune: An Efficient Data Composition Framework for Training Multi-Capability LLMs

1 code implementation18 Nov 2024 Keer Lu, Keshi Zhao, Zheng Liang, Da Pan, Shusen Zhang, Xin Wu, WeiPeng Chen, Zenan Zhou, Guosheng Dong, Bin Cui, Wentao Zhang

Despite their potential, existing work mainly focuses on domain-specific enhancements during fine-tuning, the challenge of which lies in catastrophic forgetting of knowledge across other domains.

Baichuan Alignment Technical Report

no code implementations19 Oct 2024 MingAn Lin, Fan Yang, Yanjun Shen, Haoze Sun, Tianpeng Li, Chenzheng Zhu, Tao Zhang, Miao Zheng, Xu Li, Yijie Zhou, Mingyang Chen, Yanzhao Qin, Youquan Li, Hao Liang, Fei Li, Yadong Li, Mang Wang, Guosheng Dong, Kun Fang, Jianhua Xu, Bin Cui, Wentao Zhang, Zenan Zhou, WeiPeng Chen

Baichuan-Instruct is an internal model, while Qwen2-Nova-72B and Llama3-PBM-Nova-70B are instruct versions of the Qwen2-72B and Llama-3-70B base models, optimized through Baichuan Alignment.

Facilitating Multi-turn Function Calling for LLMs via Compositional Instruction Tuning

1 code implementation16 Oct 2024 Mingyang Chen, Haoze Sun, Tianpeng Li, Fan Yang, Hao Liang, Keer Lu, Bin Cui, Wentao Zhang, Zenan Zhou, WeiPeng Chen

While current research on function calling by LLMs primarily focuses on single-turn interactions, this paper addresses the overlooked necessity for LLMs to engage in multi-turn function calling--critical for handling compositional, real-world queries that require planning with functions but not only use functions.

8k

Baichuan-Omni Technical Report

2 code implementations11 Oct 2024 Yadong Li, Haoze Sun, MingAn Lin, Tianpeng Li, Guosheng Dong, Bowen Ding, Wei Song, Zhenglin Cheng, Yuqi Huo, Song Chen, Xu Li, Da Pan, Shusen Zhang, Xin Wu, Zheng Liang, Jun Liu, Tao Zhang, Keer Lu, Yaqi Zhao, Yanjun Shen, Fan Yang, Kaicheng Yu, Tao Lin, Jianhua Xu, Zenan Zhou, WeiPeng Chen

The salient multimodal capabilities and interactive experience of GPT-4o highlight its critical role in practical applications, yet it lacks a high-performing open-source counterpart.

Language Modeling Language Modelling +3

Data Proportion Detection for Optimized Data Management for Large Language Models

1 code implementation26 Sep 2024 Hao Liang, Keshi Zhao, Yajie Yang, Bin Cui, Guosheng Dong, Zenan Zhou, Wentao Zhang

Large language models (LLMs) have demonstrated exceptional performance across a wide range of tasks and domains, with data preparation playing a critical role in achieving these results.

Management

DataSculpt: Crafting Data Landscapes for Long-Context LLMs through Multi-Objective Partitioning

2 code implementations2 Sep 2024 Keer Lu, Xiaonan Nie, Zheng Liang, Da Pan, Shusen Zhang, Keshi Zhao, WeiPeng Chen, Zenan Zhou, Guosheng Dong, Bin Cui, Wentao Zhang

Through extensive experimental analysis, we identified three key challenges in designing effective data management strategies that enable the model to achieve long-context capability without sacrificing performance in other tasks: (1) a shortage of long documents across multiple domains, (2) effective construction of context windows, and (3) efficient organization of large-scale datasets.

Code Completion Combinatorial Optimization +5

SysBench: Can Large Language Models Follow System Messages?

1 code implementation20 Aug 2024 Yanzhao Qin, Tao Zhang, Yanjun Shen, Wenjing Luo, Haoze Sun, Yan Zhang, Yujing Qiao, WeiPeng Chen, Zenan Zhou, Wentao Zhang, Bin Cui

Finally, we conduct extensive evaluation across various existing LLMs, measuring their ability to follow specified constraints given in system messages.

CFBench: A Comprehensive Constraints-Following Benchmark for LLMs

1 code implementation2 Aug 2024 Yanjun Shen, Wenjing Luo, Yan Zhang, Hao Liang, Tao Zhang, Fan Yang, MingAn Lin, Yujing Qiao, WeiPeng Chen, Bin Cui, Wentao Zhang, Zenan Zhou

The adeptness of Large Language Models (LLMs) in comprehending and following natural language instructions is critical for their deployment in sophisticated real-world applications.

PAS: Data-Efficient Plug-and-Play Prompt Augmentation System

no code implementations8 Jul 2024 Miao Zheng, Hao Liang, Fan Yang, Haoze Sun, Tianpeng Li, Lingchu Xiong, Yan Zhang, Youzhen Wu, Kun Li, Yanjun Shen, MingAn Lin, Tao Zhang, Guosheng Dong, Yujing Qiao, Kun Fang, WeiPeng Chen, Bin Cui, Wentao Zhang, Zenan Zhou

This combination of high performance, efficiency, and flexibility makes PAS a valuable system for enhancing the usability and effectiveness of LLMs through improved prompt engineering.

Prompt Engineering

Cannot find the paper you are looking for? You can Submit a new open access paper.