Search Results for author: Xinya Wu

Found 4 papers, 4 papers with code

Aquila2 Technical Report

2 code implementations14 Aug 2024 Bo-Wen Zhang, Liangdong Wang, Jijie Li, Shuhao Gu, Xinya Wu, Zhengduo Zhang, Boyan Gao, Yulong Ao, Guang Liu

This paper introduces the Aquila2 series, which comprises a wide range of bilingual models with parameter sizes of 7, 34, and 70 billion.

Management

AquilaMoE: Efficient Training for MoE Models with Scale-Up and Scale-Out Strategies

1 code implementation13 Aug 2024 Bo-Wen Zhang, Liangdong Wang, Ye Yuan, Jijie Li, Shuhao Gu, Mengdi Zhao, Xinya Wu, Guang Liu, ChengWei Wu, Hanyu Zhao, Li Du, Yiming Ju, Quanyue Ma, Yulong Ao, Yingli Zhao, Songhe Zhu, Zhou Cao, Dong Liang, Yonghua Lin, Ming Zhang, Shunfei Wang, Yanxin Zhou, Min Ye, Xuekai Chen, Xinyang Yu, Xiangjun Huang, Jian Yang

In this paper, we present AquilaMoE, a cutting-edge bilingual 8*16B Mixture of Experts (MoE) language model that has 8 experts with 16 billion parameters each and is developed using an innovative training methodology called EfficientScale.

Language Modelling Transfer Learning

CMMU: A Benchmark for Chinese Multi-modal Multi-type Question Understanding and Reasoning

1 code implementation25 Jan 2024 Zheqi He, Xinya Wu, Pengfei Zhou, Richeng Xuan, Guang Liu, Xi Yang, Qiannan Zhu, Hua Huang

However, the mastery of domain-specific knowledge, which is essential for evaluating the intelligence of MLLMs, continues to be a challenge.

Multiple-choice Position

AltDiffusion: A Multilingual Text-to-Image Diffusion Model

1 code implementation19 Aug 2023 Fulong Ye, Guang Liu, Xinya Wu, Ledell Wu

Specifically, we first train a multilingual text encoder based on the knowledge distillation.

Blocking Concept Alignment +1

Cannot find the paper you are looking for? You can Submit a new open access paper.