no code implementations • 13 Feb 2024 • Shaeke Salman, Md Montasir Bin Shams, Xiuwen Liu, Lingjiong Zhu
Transformer-based models have dominated natural language processing and other areas in the last few years due to their superior (zero-shot) performance on benchmark datasets.
no code implementations • 28 Jan 2024 • Shaeke Salman, Md Montasir Bin Shams, Xiuwen Liu
Pre-trained large foundation models play a central role in the recent surge of artificial intelligence, resulting in fine-tuned models with remarkable abilities when measured on benchmark datasets, standard exams, and applications.
no code implementations • 25 Jan 2023 • Luis Villamil, Ryan Bausback, Shaeke Salman, Ting L. Liu, Conrad Horn, Xiuwen Liu
We further incorporate weight decay, batch normalization, dropout, and label smoothing to improve the generalization of the trained models.
no code implementations • 18 Oct 2019 • Shaeke Salman, Canlin Zhang, Xiuwen Liu, Washington Mio
We show that the generalization intervals of a ReLU network behave similarly along pairwise directions between samples of the same label in both real and random cases on the MNIST and CIFAR-10 datasets.
no code implementations • 14 May 2019 • Shaeke Salman, Seyedeh Neelufar Payrovnaziri, Xiuwen Liu, Pablo Rengifo-Moreno, Zhe He
In particular, while the proposed method maintains similar interpretability as conventional shallow models such as logistic regression, it improves the prediction accuracy significantly.
no code implementations • 19 Jan 2019 • Shaeke Salman, Xiuwen Liu
Assisted by the availability of data and high performance computing, deep learning techniques have achieved breakthroughs and surpassed human performance empirically in difficult tasks, including object recognition, speech recognition, and natural language processing.