Search Results for author: Quang Pham

Found 17 papers, 12 papers with code

CompeteSMoE - Effective Training of Sparse Mixture of Experts via Competition

no code implementations4 Feb 2024 Quang Pham, Giang Do, Huy Nguyen, TrungTin Nguyen, Chenghao Liu, Mina Sartipi, Binh T. Nguyen, Savitha Ramasamy, XiaoLi Li, Steven Hoi, Nhat Ho

Sparse mixture of experts (SMoE) offers an appealing solution to scale up the model complexity beyond the mean of increasing the network's depth or width.

HyperRouter: Towards Efficient Training and Inference of Sparse Mixture of Experts

1 code implementation12 Dec 2023 Giang Do, Khiem Le, Quang Pham, TrungTin Nguyen, Thanh-Nam Doan, Bint T. Nguyen, Chenghao Liu, Savitha Ramasamy, XiaoLi Li, Steven Hoi

By routing input tokens to only a few split experts, Sparse Mixture-of-Experts has enabled efficient training of large language models.

Adaptive-saturated RNN: Remember more with less instability

1 code implementation ICLR 2023 Tiny Paper Track 2023 Khoi Minh Nguyen-Duy, Quang Pham, Binh T. Nguyen

Orthogonal parameterization is a compelling solution to the vanishing gradient problem (VGP) in recurrent neural networks (RNNs).

Continual Learning, Fast and Slow

1 code implementation6 Sep 2022 Quang Pham, Chenghao Liu, Steven C. H. Hoi

Motivated by this theory, we propose \emph{DualNets} (for Dual Networks), a general continual learning framework comprising a fast learning system for supervised learning of pattern-separated representation from specific tasks and a slow learning system for representation learning of task-agnostic general representation via Self-Supervised Learning (SSL).

Continual Learning Hippocampus +2

Continual Normalization: Rethinking Batch Normalization for Online Continual Learning

1 code implementation ICLR 2022 Quang Pham, Chenghao Liu, Steven Hoi

Existing continual learning methods use Batch Normalization (BN) to facilitate training and improve generalization across tasks.

Continual Learning

Learning Fast and Slow for Online Time Series Forecasting

1 code implementation23 Feb 2022 Quang Pham, Chenghao Liu, Doyen Sahoo, Steven C. H. Hoi

The fast adaptation capability of deep neural networks in non-stationary environments is critical for online time series forecasting.

Time Series Time Series Forecasting

DualNet: Continual Learning, Fast and Slow

1 code implementation NeurIPS 2021 Quang Pham, Chenghao Liu, Steven Hoi

According to Complementary Learning Systems (CLS) theory~\citep{mcclelland1995there} in neuroscience, humans do effective \emph{continual learning} through two complementary systems: a fast learning system centered on the hippocampus for rapid learning of the specifics and individual experiences, and a slow learning system located in the neocortex for the gradual acquisition of structured knowledge about the environment.

Continual Learning Hippocampus +2

An Efficient Transformer-Based Model for Vietnamese Punctuation Prediction

1 code implementation IEA/AIE 2021 Hieu Tran, Cuong V. Dinh, Quang Pham, Binh T. Nguyen

In both formal and informal texts, missing punctuation marks make the texts confusing and challenging to read.

TATL: Task Agnostic Transfer Learning for Skin Attributes Detection

no code implementations4 Apr 2021 Duy M. H. Nguyen, Thu T. Nguyen, Huong Vu, Quang Pham, Manh-Duy Nguyen, Binh T. Nguyen, Daniel Sonntag

Existing skin attributes detection methods usually initialize with a pre-trained Imagenet network and then fine-tune on a medical target task.

Attribute Transfer Learning

Contextual Transformation Networks for Online Continual Learning

no code implementations ICLR 2021 Quang Pham, Chenghao Liu, Doyen Sahoo, Steven Hoi

Continual learning methods with fixed architectures rely on a single network to learn models that can perform well on all tasks.

Continual Learning Transfer Learning

Online Continual Learning Under Domain Shift

no code implementations1 Jan 2021 Quang Pham, Chenghao Liu, Steven Hoi

CIER employs an adversarial training to correct the shift in $P(X, Y)$ by matching $P(X|Y)$, which results in an invariant representation that can generalize to unseen domains during inference.

Continual Learning

Bilevel Continual Learning

1 code implementation30 Jul 2020 Quang Pham, Doyen Sahoo, Chenghao Liu, Steven C. H. Hoi

Continual learning aims to learn continuously from a stream of tasks and data in an online-learning fashion, being capable of exploiting what was learned previously to improve current and future tasks while still being able to perform well on the previous tasks.

Bilevel Optimization Continual Learning +2

Extracting Entities and Topics from News and Connecting Criminal Records

1 code implementation3 May 2020 Quang Pham, Marija Stanojevic, Zoran Obradovic

The goal of this paper is to summarize methodologies used in extracting entities and topics from a database of criminal records and from a database of newspapers.

Clustering

URLNet: Learning a URL Representation with Deep Learning for Malicious URL Detection

3 code implementations9 Feb 2018 Hung Le, Quang Pham, Doyen Sahoo, Steven C. H. Hoi

This approach allows the model to capture several types of semantic information, which was not possible by the existing models.

BIG-bench Machine Learning Feature Engineering +1

Online Deep Learning: Learning Deep Neural Networks on the Fly

4 code implementations10 Nov 2017 Doyen Sahoo, Quang Pham, Jing Lu, Steven C. H. Hoi

Deep Neural Networks (DNNs) are typically trained by backpropagation in a batch learning setting, which requires the entire training data to be made available prior to the learning task.

Cannot find the paper you are looking for? You can Submit a new open access paper.