no code implementations • 15 Apr 2024 • Yang Lin, Xinyu Ma, Xu Chu, Yujie Jin, Zhibang Yang, Yasha Wang, Hong Mei
We then demonstrate the theoretical mechanism of our LoRA Dropout mechanism from the perspective of sparsity regularization by providing a generalization error bound under this framework.
no code implementations • 5 Apr 2024 • Xinyu Ma, Xu Chu, Zhibang Yang, Yang Lin, Xin Gao, Junfeng Zhao
With the increasingly powerful performances and enormous scales of Pretrained Language Models (PLMs), promoting parameter efficiency in fine-tuning has become a crucial need for effective and efficient adaptation to various downstream tasks.
no code implementations • 13 Dec 2022 • Qinyi Deng, Yong Guo, Zhibang Yang, Haolin Pan, Jian Chen
In this way, these data can be also very informative if we can effectively exploit these complementary labels, i. e., the classes that a sample does not belong to.