Search Results for author: Jun Zhan

Found 8 papers, 6 papers with code

Data Mixing Laws: Optimizing Data Mixtures by Predicting Language Modeling Performance

1 code implementation25 Mar 2024 Jiasheng Ye, Peiju Liu, Tianxiang Sun, Yunhua Zhou, Jun Zhan, Xipeng Qiu

Pretraining data of large language models composes multiple domains (e. g., web texts, academic papers, codes), whose mixture proportions crucially impact the competence of outcome models.

Language Modelling

AnyGPT: Unified Multimodal LLM with Discrete Sequence Modeling

1 code implementation19 Feb 2024 Jun Zhan, Junqi Dai, Jiasheng Ye, Yunhua Zhou, Dong Zhang, Zhigeng Liu, Xin Zhang, Ruibin Yuan, Ge Zhang, Linyang Li, Hang Yan, Jie Fu, Tao Gui, Tianxiang Sun, Yugang Jiang, Xipeng Qiu

We introduce AnyGPT, an any-to-any multimodal language model that utilizes discrete representations for the unified processing of various modalities, including speech, text, images, and music.

Language Modelling Large Language Model

SpeechGPT-Gen: Scaling Chain-of-Information Speech Generation

1 code implementation24 Jan 2024 Dong Zhang, Xin Zhang, Jun Zhan, ShiMin Li, Yaqian Zhou, Xipeng Qiu

It comprises an autoregressive model based on LLM for semantic information modeling and a non-autoregressive model employing flow matching for perceptual information modeling.

Voice Conversion

SpeechGPT: Empowering Large Language Models with Intrinsic Cross-Modal Conversational Abilities

1 code implementation18 May 2023 Dong Zhang, ShiMin Li, Xin Zhang, Jun Zhan, Pengyu Wang, Yaqian Zhou, Xipeng Qiu

Multi-modal large language models are regarded as a crucial step towards Artificial General Intelligence (AGI) and have garnered significant interest with the emergence of ChatGPT.

Language Modelling Large Language Model +2

HFN: Heterogeneous Feature Network for Multivariate Time Series Anomaly Detection

no code implementations1 Nov 2022 Jun Zhan, Chengkun Wu, Canqun Yang, Qiucheng Miao, Xiandong Ma

In this paper, we propose a novel semi-supervised anomaly detection framework based on a heterogeneous feature network (HFN) for MTS, learning heterogeneous structure information from a mass of unlabeled time-series data to improve the accuracy of anomaly detection, and using attention coefficient to provide an explanation for the detected anomalies.

Graph structure learning Representation Learning +4

Aerodynamic Data Predictions Based on Multi-task Learning

no code implementations15 Oct 2020 Liwei Hu, Yu Xiang, Jun Zhan, Zifang Shi, Wenzheng Wang

Predicting high-speed data is more difficult than predicting low-speed data, owing to that the number of high-speed data is limited, i. e. the quality of the Burgers' dataset is not satisfactory.

Multi-Task Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.