1 code implementation • 6 Feb 2024 • Zhanpeng Zhou, Zijun Chen, Yilan Chen, Bo Zhang, Junchi Yan
The pretraining-finetuning paradigm has become the prevailing trend in modern deep learning.
no code implementations • 9 Oct 2023 • Justin Lee, Tuomas Oikarinen, Arjun Chatha, Keng-Chi Chang, Yilan Chen, Tsui-Wei Weng
Recent advances have greatly increased the capabilities of large language models (LLMs), but our understanding of the models and their safety has not progressed as fast.
no code implementations • 18 Aug 2022 • Quanshi Zhang, Xu Cheng, Yilan Chen, Zhefan Rao
This paper provides a new perspective to explain the success of knowledge distillation, i. e., quantifying knowledge points encoded in intermediate layers of a DNN for classification, based on the information theory.
no code implementations • 4 Feb 2022 • Wei Huang, Chunrui Liu, Yilan Chen, Tianyu Liu, Richard Yi Da Xu
In addition to being a pure generalization bound analysis tool, PAC-Bayesian bound can also be incorporated into an objective function to train a probabilistic neural network, making them a powerful and relevant framework that can numerically provide a tight generalization bound for supervised learning.
1 code implementation • NeurIPS 2021 • Yilan Chen, Wei Huang, Lam M. Nguyen, Tsui-Wei Weng
Therefore, in this work, we propose to establish the equivalence between NN and SVM, and specifically, the infinitely wide NN trained by soft margin loss and the standard soft margin SVM with NTK trained by subgradient descent.
no code implementations • CVPR 2020 • Xu Cheng, Zhefan Rao, Yilan Chen, Quanshi Zhang
Whereas, in the scenario of learning from raw data, the DNN learns visual concepts sequentially.