Search Results for author: Chaoqi Liang

Found 2 papers, 0 papers with code

IMWA: Iterative Model Weight Averaging Benefits Class-Imbalanced Learning Tasks

no code implementations25 Apr 2024 Zitong Huang, Ze Chen, Bowen Dong, Chaoqi Liang, Erjin Zhou, WangMeng Zuo

Model Weight Averaging (MWA) is a technique that seeks to enhance model's performance by averaging the weights of multiple trained models.

Rethinking the BERT-like Pretraining for DNA Sequences

no code implementations11 Oct 2023 Chaoqi Liang, Weiqiang Bai, Lifeng Qiao, Yuchen Ren, Jianle Sun, Peng Ye, Hongliang Yan, Xinzhu Ma, WangMeng Zuo, Wanli Ouyang

To address this research gap, we first conducted a series of exploratory experiments and gained several insightful observations: 1) In the fine-tuning phase of downstream tasks, when using K-mer overlapping tokenization instead of K-mer non-overlapping tokenization, both overlapping and non-overlapping pretraining weights show consistent performance improvement. 2) During the pre-training process, using K-mer overlapping tokenization quickly produces clear K-mer embeddings and reduces the loss to a very low level, while using K-mer non-overlapping tokenization results in less distinct embeddings and continuously decreases the loss.

Cannot find the paper you are looking for? You can Submit a new open access paper.