1 code implementation • 31 Aug 2024 • Zhiyuan Hu, Yuliang Liu, Jinman Zhao, Suyuchen Wang, Yan Wang, Wei Shen, Qing Gu, Anh Tuan Luu, See-Kiong Ng, Zhiwei Jiang, Bryan Hooi
Large language models (LLMs) face significant challenges in handling long-context tasks because of their limited effective context window size during pretraining, which restricts their ability to generalize over extended sequences.
no code implementations • 25 Jun 2024 • Do Huu Dat, Do Duc Anh, Anh Tuan Luu, Wray Buntine
While diffusion models excel at conditional generating high-quality images, prior works in discrete diffusion models were not evaluated on conditional long-text generation.
no code implementations • 30 May 2024 • Thong Thanh Nguyen, Zhiyuan Hu, Xiaobao Wu, Cong-Duy T Nguyen, See-Kiong Ng, Anh Tuan Luu
Seeking answers effectively for long videos is essential to build video question answering (videoQA) systems.
2 code implementations • 28 May 2024 • Xiaobao Wu, Xinshuai Dong, Liangming Pan, Thong Nguyen, Anh Tuan Luu
However, existing models suffer from repetitive topic and unassociated topic issues, failing to reveal the evolution and hindering further applications.
2 code implementations • 28 May 2024 • Xiaobao Wu, Thong Nguyen, Delvin Ce Zhang, William Yang Wang, Anh Tuan Luu
This brings about a neat and efficient topic modeling framework.
1 code implementation • 26 Mar 2024 • Cong-Duy Nguyen, Thong Nguyen, Xiaobao Wu, Anh Tuan Luu
Previous work on multimodal sentence embedding has proposed multimodal contrastive learning and achieved promising results.
no code implementations • 15 Mar 2024 • Chaoqun Liu, Wenxuan Zhang, Yiran Zhao, Anh Tuan Luu, Lidong Bing
Large language models (LLMs) have demonstrated multilingual capabilities; yet, they are mostly English-centric due to the imbalanced training corpora.
no code implementations • 5 Mar 2024 • Bosheng Ding, Chengwei Qin, Ruochen Zhao, Tianze Luo, Xinze Li, Guizhen Chen, Wenhan Xia, Junjie Hu, Anh Tuan Luu, Shafiq Joty
In the rapidly evolving field of large language models (LLMs), data augmentation (DA) has emerged as a pivotal technique for enhancing model performance by diversifying training examples without the need for additional data collection.
1 code implementation • 29 Feb 2024 • Xiaobao Wu, Liangming Pan, William Yang Wang, Anh Tuan Luu
In this paper, we propose a new benchmark, Unstructured Knowledge Editing (UKE).
1 code implementation • 25 Feb 2024 • Xin Mao, Feng-Lin Li, Huimin Xu, Wei zhang, Anh Tuan Luu
While Reinforcement Learning from Human Feedback (RLHF) significantly enhances the generation quality of Large Language Models (LLMs), recent studies have raised concerns regarding the complexity and instability associated with the Proximal Policy Optimization (PPO) algorithm, proposing a series of order-based calibration methods as viable alternatives.
no code implementations • 12 Feb 2024 • Thong Nguyen, Xiaobao Wu, Xinshuai Dong, Cong-Duy T Nguyen, See-Kiong Ng, Anh Tuan Luu
Secondly, we explicitly cast contrastive topic modeling as a gradient-based multi-objective optimization problem, with the goal of achieving a Pareto stationary solution that balances the trade-off between the ELBO and the contrastive objective.
1 code implementation • 12 Feb 2024 • Mingzhe Du, Anh Tuan Luu, Bin Ji, Qian Liu, See-Kiong Ng
Based on the distribution, we introduce a new metric Beyond, which computes a runtime-percentile-weighted Pass score to reflect functional correctness and code efficiency simultaneously.
1 code implementation • 5 Feb 2024 • Zhiyuan Hu, Chumin Liu, Xidong Feng, Yilun Zhao, See-Kiong Ng, Anh Tuan Luu, Junxian He, Pang Wei Koh, Bryan Hooi
In the face of uncertainty, the ability to *seek information* is of fundamental importance.
2 code implementations • 27 Jan 2024 • Xiaobao Wu, Thong Nguyen, Anh Tuan Luu
In this paper, we present a comprehensive survey on neural topic models concerning methods, applications, and challenges.
2 code implementations • 25 Jan 2024 • Xiaobao Wu, Fengjun Pan, Thong Nguyen, Yichao Feng, Chaoqun Liu, Cong-Duy Nguyen, Anh Tuan Luu
Hierarchical topic modeling aims to discover latent topics from a corpus and organize them into a hierarchy to understand documents with desirable semantic granularity.
1 code implementation • 9 Jan 2024 • Khoi M. Le, Trinh Pham, Tho Quan, Anh Tuan Luu
Paraphrases are texts that convey the same meaning while using different words or sentence structures.
1 code implementation • 18 Dec 2023 • Vijay Prakash Dwivedi, Yozen Liu, Anh Tuan Luu, Xavier Bresson, Neil Shah, Tong Zhao
As such, a key innovation of this work lies in the creation of a fast neighborhood sampling technique coupled with a local attention mechanism that encompasses a 4-hop reception field, but achieved through just 2-hop operations.
1 code implementation • 4 Dec 2023 • Phuoc Pham Van Long, Duc Anh Vu, Nhat M. Hoang, Xuan Long Do, Anh Tuan Luu
In the context-unaware setting, we evaluate ChatGPT in generating math questions for each lesson from pre-university math curriculums that we crawl.
1 code implementation • 22 Oct 2023 • Mingzhe Du, Anh Tuan Luu, Bin Ji, See-Kiong Ng
The vast number of parameters in large language models (LLMs) endows them with remarkable capabilities, allowing them to excel in a variety of natural language processing tasks.
no code implementations • 16 Sep 2023 • Zhiyuan Hu, Yue Feng, Yang Deng, Zekun Li, See-Kiong Ng, Anh Tuan Luu, Bryan Hooi
Recently, the development of large language models (LLMs) has been significantly enhanced the question answering and dialogue generation, and makes them become increasingly popular in current practical scenarios.
1 code implementation • 13 Sep 2023 • Xiaobao Wu, Fengjun Pan, Anh Tuan Luu
Topic models have a rich history with various applications and have recently been reinvigorated by neural topic modeling.
1 code implementation • 3 Sep 2023 • Yue Zhang, Yafu Li, Leyang Cui, Deng Cai, Lemao Liu, Tingchen Fu, Xinting Huang, Enbo Zhao, Yu Zhang, Yulong Chen, Longyue Wang, Anh Tuan Luu, Wei Bi, Freda Shi, Shuming Shi
While large language models (LLMs) have demonstrated remarkable capabilities across a range of downstream tasks, a significant concern revolves around their propensity to exhibit hallucinations: LLMs occasionally generate content that diverges from the user input, contradicts previously generated context, or misaligns with established world knowledge.
1 code implementation • 16 Jun 2023 • Zhiyuan Hu, Yue Feng, Anh Tuan Luu, Bryan Hooi, Aldo Lipani
This approach uses LLM as annotation-free user simulator to assess dialogue responses, combining them with smaller fine-tuned end-to-end TOD models.
1 code implementation • 14 Jun 2023 • Zhiyuan Hu, Chumin Liu, Yue Feng, Anh Tuan Luu, Bryan Hooi
Controllable text generation is a challenging and meaningful field in natural language generation (NLG).
2 code implementations • 7 Jun 2023 • Xiaobao Wu, Xinshuai Dong, Thong Nguyen, Anh Tuan Luu
Topic models have been prevalent for decades with various applications.
no code implementations • 25 May 2023 • Yandan Zheng, Anran Hao, Anh Tuan Luu
To alleviate the issues, we propose Jointprop, a Heterogeneous Graph-based Propagation framework for joint semi-supervised entity and relation extraction, which captures the global structure information between individual tasks and exploits interactions within unlabeled data.
1 code implementation • 22 May 2023 • Liangming Pan, Xiaobao Wu, Xinyuan Lu, Anh Tuan Luu, William Yang Wang, Min-Yen Kan, Preslav Nakov
Fact-checking real-world claims often requires collecting multiple pieces of evidence and applying complex multi-step reasoning.
1 code implementation • 22 May 2023 • Thong Nguyen, Xiaobao Wu, Xinshuai Dong, Anh Tuan Luu, Cong-Duy Nguyen, Zhen Hai, Lidong Bing
Multimodal Review Helpfulness Prediction (MRHP) aims to rank product reviews based on predicted helpfulness scores and has been widely applied in e-commerce via presenting customers with useful reviews.
1 code implementation • 19 May 2023 • Chaoqun Liu, Wenxuan Zhang, Guizhen Chen, Xiaobao Wu, Anh Tuan Luu, Chip Hong Chang, Lidong Bing
In this work, we propose a new paradigm based on self-supervised learning to solve zero-shot text classification tasks by tuning the language models with unlabeled data, called self-supervised tuning.
2 code implementations • 7 Apr 2023 • Xiaobao Wu, Xinshuai Dong, Thong Nguyen, Chaoqun Liu, Liangming Pan, Anh Tuan Luu
Instead of the direct alignment in previous work, we propose a topic alignment with mutual information method.
2 code implementations • 23 Nov 2022 • Xiaobao Wu, Anh Tuan Luu, Xinshuai Dong
To overcome the data sparsity issue in short text topic modeling, existing methods commonly rely on data augmentation or the data characteristic of short texts to introduce more word co-occurrence information.
1 code implementation • 18 Nov 2022 • Leon O. Guertler, Andri Ashfahani, Anh Tuan Luu
The long-standing challenge of building effective classification models for small and imbalanced datasets has seen little improvement since the creation of the Synthetic Minority Over-sampling Technique (SMOTE) over 20 years ago.
1 code implementation • 5 Jul 2022 • Thong Nguyen, Cong-Duy Nguyen, Xiaobao Wu, See-Kiong Ng, Anh Tuan Luu
Moreover, a list of training datasets and downstream tasks is supplied to further polish the perspective into V\&L pretraining.
2 code implementations • 16 Jun 2022 • Vijay Prakash Dwivedi, Ladislav Rampášek, Mikhail Galkin, Ali Parviz, Guy Wolf, Anh Tuan Luu, Dominique Beaini
Graph Neural Networks (GNNs) that are based on the message passing (MP) paradigm generally exchange information between 1-hop neighbors to build node representations at each layer.
Ranked #3 on Link Prediction on PCQM-Contact
3 code implementations • 25 May 2022 • Ladislav Rampášek, Mikhail Galkin, Vijay Prakash Dwivedi, Anh Tuan Luu, Guy Wolf, Dominique Beaini
We propose a recipe on how to build a general, powerful, scalable (GPS) graph Transformer with linear complexity and state-of-the-art results on a diverse set of benchmarks.
Ranked #1 on Graph Property Prediction on ogbg-ppa
1 code implementation • 24 May 2022 • Haiteng Zhao, Chang Ma, Xinshuai Dong, Anh Tuan Luu, Zhi-Hong Deng, Hanwang Zhang
Deep learning models have achieved great success in many fields, yet they are vulnerable to adversarial examples.
2 code implementations • NeurIPS 2021 • Thong Nguyen, Anh Tuan Luu
Recent empirical studies show that adversarial topic models (ATM) can successfully capture semantic patterns of the document by differentiating a document with another dissimilar sample.
1 code implementation • ICLR 2022 • Vijay Prakash Dwivedi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, Xavier Bresson
An approach to tackle this issue is to introduce Positional Encoding (PE) of nodes, and inject it into the input layer, like in Transformers.
Ranked #14 on Graph Regression on ZINC-500k
no code implementations • 29 Sep 2021 • Cong-Duy T Nguyen, Anh Tuan Luu, Tho Quan
However, this approach has two main drawbacks: (i) the whole image usually contains more objects and backgrounds than the sentence itself; thus, matching them together will confuse the grounded model; (ii) CNN only extracts the features of the image but not the relationship between objects inside that, limiting the grounded model to learn complicated contexts.
no code implementations • EMNLP 2021 • Thong Nguyen, Anh Tuan Luu, Truc Lu, Tho Quan
Recently, Transformer-based models have been proven effective in the abstractive summarization task by creating fluent and informative summaries.
1 code implementation • ICLR 2021 • Xinshuai Dong, Anh Tuan Luu, Rongrong Ji, Hong Liu
Robustness against word substitutions has a well-defined and widely acceptable form, i. e., using semantically similar words as substitutions, and thus it is considered as a fundamental stepping-stone towards broader robustness in natural language processing.
3 code implementations • 17 Feb 2021 • Aston Zhang, Yi Tay, Shuai Zhang, Alvin Chan, Anh Tuan Luu, Siu Cheung Hui, Jie Fu
Recent works have demonstrated reasonable success of representation learning in hypercomplex space.
no code implementations • ICLR 2021 • Aston Zhang, Yi Tay, Shuai Zhang, Alvin Chan, Anh Tuan Luu, Siu Hui, Jie Fu
Recent works have demonstrated reasonable success of representation learning in hypercomplex space.
no code implementations • ACL 2020 • Yi Tay, Donovan Ong, Jie Fu, Alvin Chan, Nancy Chen, Anh Tuan Luu, Chris Pal
Understanding human preferences, along with cultural and social nuances, lives at the heart of natural language understanding.
16 code implementations • 2 Mar 2020 • Vijay Prakash Dwivedi, Chaitanya K. Joshi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, Xavier Bresson
In the last few years, graph neural networks (GNNs) have become the standard toolkit for analyzing and learning from data on graphs.
Ranked #1 on Link Prediction on COLLAB
no code implementations • NeurIPS 2019 • Yi Tay, Anh Tuan Luu, Aston Zhang, Shuohang Wang, Siu Cheung Hui
Attentional models are distinctly characterized by their ability to learn relative importance, i. e., assigning a different weight to input values.
1 code implementation • AAAI 2019 • Yi Tay, Shuai Zhang, Anh Tuan Luu, Siu Cheung Hui, Lina Yao, Tran Dang Quang Vinh
Factorization Machines (FMs) are a class of popular algorithms that have been widely adopted for collaborative filtering and recommendation tasks.
no code implementations • 12 Nov 2018 • Anran Wang, Anh Tuan Luu, Chuan-Sheng Foo, Hongyuan Zhu, Yi Tay, Vijay Chandrasekhar
In this paper, we present the Holistic Multi-modal Memory Network (HMMN) framework which fully considers the interactions between different input sources (multi-modal context, question) in each hop.
no code implementations • EMNLP 2018 • Yi Tay, Anh Tuan Luu, Siu Cheung Hui, Jian Su
This paper proposes a new neural architecture that exploits readily available sentiment lexicon resources.
no code implementations • EMNLP 2018 • Yi Tay, Anh Tuan Luu, Siu Cheung Hui
Sequence encoders are crucial components in many neural architectures for learning to read and comprehend.
Ranked #7 on Question Answering on NarrativeQA
no code implementations • 29 May 2018 • Yi Tay, Anh Tuan Luu, Siu Cheung Hui
Our approach, the CoupleNet is an end-to-end deep learning based estimator that analyzes the social profiles of two users and subsequently performs a similarity match between the users.
no code implementations • 14 Dec 2017 • Yi Tay, Anh Tuan Luu, Siu Cheung Hui
Our novel model, \textit{Aspect Fusion LSTM} (AF-LSTM) learns to attend based on associative relationships between sentence words and aspect which allows our model to adaptively focus on the correct words given an aspect term.
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +1
1 code implementation • 17 Jul 2017 • Yi Tay, Anh Tuan Luu, Siu Cheung Hui
Our model, LRML (\textit{Latent Relational Metric Learning}) is a novel metric learning approach for recommendation.
Ranked #1 on Recommendation Systems on Netflix (nDCG@10 metric)