no code implementations • NAACL 2022 • Hung Le, Nancy Chen, Steven Hoi
Neural module networks (NMN) have achieved success in image-grounded tasks such as Visual Question Answering (VQA) on synthetic images.
no code implementations • 24 Aug 2024 • Bao Duong, Hung Le, Thin Nguyen
Recently, reinforcement learning (RL) has proved a promising alternative for conventional local heuristics in score-based approaches to learning directed acyclic causal graphs (DAGs) from observational data.
no code implementations • 14 Aug 2024 • Dai Do, Quan Tran, Svetha Venkatesh, Hung Le
We approach prompt optimization as a Reinforcement Learning (RL) challenge, using episodic memory to archive combinations of input data, permutations of few-shot examples, and the rewards observed during training.
1 code implementation • 17 Jul 2024 • Minh Hoang Nguyen, Hung Le, Svetha Venkatesh
In this paper, we introduce a novel framework, Variable-Agnostic Causal Exploration for Reinforcement Learning (VACERL), incorporating causal relationships to drive exploration in RL without specifying environmental causal variables.
no code implementations • 9 Jul 2024 • Thanh-Dat Nguyen, Tung Do-Viet, Hung Nguyen-Duy, Tuan-Hai Luu, Hung Le, Bach Le, Patanamon, Thongtanunam
VRDSynth(Table) outperforms these baselines in 4 out of 8 languages and in average F1 score.
no code implementations • 23 Jun 2024 • Hung Le, Yingbo Zhou, Caiming Xiong, Silvio Savarese, Doyen Sahoo
In this work, we introduce INDICT: a new framework that empowers LLMs with Internal Dialogues of Critiques for both safety and helpfulness guidance.
no code implementations • 22 Jun 2024 • Giang Do, Hung Le, Truyen Tran
Sparse mixture of experts (SMoE) have emerged as an effective approach for scaling large language models while keeping a constant computational cost.
no code implementations • 26 May 2024 • Hung Le, Quan Tran, Dung Nguyen, Kien Do, Saloni Mittal, Kelechi Ogueji, Svetha Venkatesh
Recent approaches, such as direct preference optimization (DPO), have eliminated the need for unstable and sluggish reinforcement learning optimization by introducing close-formed supervised losses.
no code implementations • 18 Apr 2024 • Hung Le, Dung Nguyen, Kien Do, Svetha Venkatesh, Truyen Tran
We propose Pointer-Augmented Neural Memory (PANM) to help neural networks understand and apply symbol processing to new, longer sequences of data.
no code implementations • 3 Apr 2024 • Viet-Tung Do, Van-Khanh Hoang, Duy-Hung Nguyen, Shahab Sabahi, Jeff Yang, Hajime Hotta, Minh-Tien Nguyen, Hung Le
Our approach consists of three steps: (1) clustering the training data and generating candidate prompts for each cluster using an LLM-based prompt generator; (2) synthesizing a dataset of input-prompt-output tuples for training a prompt evaluator to rank the prompts based on their relevance to the input; (3) using the prompt evaluator to select the best prompt for a new input at test time.
1 code implementation • 5 Feb 2024 • Kien Do, Duc Kieu, Toan Nguyen, Dang Nguyen, Hung Le, Dung Nguyen, Thin Nguyen
We propose a systematic training-free method to transform the probability flow of a "linear" stochastic process characterized by the equation X_{t}=a_{t}X_{0}+\sigma_{t}X_{1} into a straight constant-speed (SC) flow, reminiscent of Rectified Flow.
no code implementations • 5 Feb 2024 • Kien Do, Dung Nguyen, Hung Le, Thao Le, Dang Nguyen, Haripriya Harikumar, Truyen Tran, Santu Rana, Svetha Venkatesh
To overcome this challenge, we propose to approximate \frac{1}{p(u|b)} using a biased classifier trained with "bias amplification" losses.
1 code implementation • 3 Jan 2024 • David Junhao Zhang, Dongxu Li, Hung Le, Mike Zheng Shou, Caiming Xiong, Doyen Sahoo
This work presents Moonshot, a new video generation model that conditions simultaneously on multimodal inputs of image and text.
1 code implementation • 16 Nov 2023 • Hung Le, Ong Eng-Jon, Bober Miroslaw
This study introduces "SurvTimeSurvival: Survival Analysis On Patients With Multiple Visits/Records", utilizing the Transformer model to not only handle the complexities of time-varying covariates but also covariates data.
1 code implementation • 13 Oct 2023 • Hung Le, Hailin Chen, Amrita Saha, Akash Gokul, Doyen Sahoo, Shafiq Joty
We find that by naturally encouraging the LLM to reuse the previously developed and verified sub-modules, CodeChain can significantly boost both modularity as well as correctness of the generated solutions, achieving relative pass@1 improvements of 35% on APPS and 76% on CodeContests.
Ranked #3 on Code Generation on CodeContests
no code implementations • 27 Aug 2023 • Thanh Duc Hoang, Do Viet Tung, Duy-Hung Nguyen, Bao-Sinh Nguyen, Huy Hoang Nguyen, Hung Le
We address catastrophic forgetting issues in graph learning as incoming data transits from one to another graph distribution.
1 code implementation • 21 Aug 2023 • Thommen George Karimpanal, Laknath Buddhika Semage, Santu Rana, Hung Le, Truyen Tran, Sunil Gupta, Svetha Venkatesh
To address this issue, we introduce SEQ (sample efficient querying), where we simultaneously train a secondary RL agent to decide when the LLM should be queried for solutions.
1 code implementation • 9 Aug 2023 • Hung Le, Kien Do, Dung Nguyen, Svetha Venkatesh
We present a new computing model for intrinsic rewards in reinforcement learning that addresses the limitations of existing surprise-driven explorations.
1 code implementation • 31 May 2023 • Nghi D. Q. Bui, Hung Le, Yue Wang, Junnan Li, Akhilesh Deepak Gotmare, Steven C. H. Hoi
In this paper, we present CodeTF, an open-source Transformer-based library for state-of-the-art Code LLMs and code intelligence.
2 code implementations • 13 May 2023 • Yue Wang, Hung Le, Akhilesh Deepak Gotmare, Nghi D. Q. Bui, Junnan Li, Steven C. H. Hoi
To address these limitations, we propose ``CodeT5+'', a family of encoder-decoder LLMs for code in which component modules can be flexibly combined to suit a wide range of downstream code tasks.
Ranked #1 on Code Search on CodeXGLUE - AdvTest
no code implementations • 12 May 2023 • Minh-Tien Nguyen, Duy-Hung Nguyen, Shahab Sabahi, Hung Le, Jeff Yang, Hajime Hotta
Based on the task we design a new model relied on LLMs which are empowered by additional knowledge extracted from insurance policy rulebooks and DBpedia.
no code implementations • 3 Mar 2023 • Sunil Gupta, Alistair Shilton, Arun Kumar A V, Shannon Ryan, Majid Abdolshah, Hung Le, Santu Rana, Julian Berk, Mahad Rashid, Svetha Venkatesh
In this paper we introduce BO-Muse, a new approach to human-AI teaming for the optimization of expensive black-box functions.
no code implementations • 17 Jan 2023 • Dung Nguyen, Phuoc Nguyen, Hung Le, Kien Do, Svetha Venkatesh, Truyen Tran
Social reasoning necessitates the capacity of theory of mind (ToM), the ability to contextualise and attribute mental states to others without having access to their internal cognitive structure.
no code implementations • 23 Oct 2022 • Kha Pham, Hung Le, Man Ngo, Truyen Tran
FINE consists of a backbone network and a trainable semantic memory of basis weight matrices.
no code implementations • 26 Sep 2022 • Bao-Sinh Nguyen, Dung Tien Le, Hieu M. Vu, Tuan Anh D. Nguyen, Minh-Tien Nguyen, Hung Le
In this paper, we investigate the problem of improving the performance of Artificial Intelligence systems in understanding document images, especially in cases where training data is limited.
no code implementations • 21 Sep 2022 • Kien Do, Hung Le, Dung Nguyen, Dang Nguyen, Haripriya Harikumar, Truyen Tran, Santu Rana, Svetha Venkatesh
Since the EMA generator can be considered as an ensemble of the generator's old versions and often undergoes a smaller change in updates compared to the generator, training on its synthetic samples can help the student recall the past knowledge and prevent the student from adapting too quickly to new updates of the generator.
1 code implementation • 15 Sep 2022 • Dongxu Li, Junnan Li, Hung Le, Guangsen Wang, Silvio Savarese, Steven C. H. Hoi
We introduce LAVIS, an open-source deep learning library for LAnguage-VISion research and applications.
2 code implementations • 5 Jul 2022 • Hung Le, Yue Wang, Akhilesh Deepak Gotmare, Silvio Savarese, Steven C. H. Hoi
To address the limitations, we propose "CodeRL", a new framework for program synthesis tasks through pretrained LMs and deep reinforcement learning (RL).
Ranked #2 on Code Generation on APPS
1 code implementation • NAACL 2022 • Hung Le, Nancy F. Chen, Steven C. H. Hoi
Specifically, we introduce a novel dialogue state tracking task to track the information of visual objects that are mentioned in video-grounded dialogues.
no code implementations • 1 Jun 2022 • Bao-Sinh Nguyen, Quang-Bach Tran, Tuan-Anh Nguyen Dang, Duc Nguyen, Hung Le
Measuring the confidence of AI models is critical for safely deploying AI in real-world industrial systems.
2 code implementations • 1 Jun 2022 • Wenzhuo Yang, Hung Le, Tanmay Laud, Silvio Savarese, Steven C. H. Hoi
We introduce OmniXAI (short for Omni eXplainable AI), an open-source Python library of eXplainable AI (XAI), which offers omni-way explainable AI capabilities and various interpretable machine learning techniques to address the pain points of understanding and interpreting the decisions made by machine learning (ML) in practice.
no code implementations • 20 Apr 2022 • Hung Le, Thommen Karimpanal George, Majid Abdolshah, Dung Nguyen, Kien Do, Sunil Gupta, Svetha Venkatesh
We introduce a constrained optimization method for policy gradient reinforcement learning, which uses a virtual trust region to regulate each policy update.
no code implementations • 17 Apr 2022 • Dung Nguyen, Phuoc Nguyen, Hung Le, Kien Do, Svetha Venkatesh, Truyen Tran
Inspired by the observation that humans often infer the character traits of others, then use it to explain behaviour, we propose a new neural ToM architecture that learns to generate a latent trait vector of an actor from the past trajectories.
no code implementations • Findings (NAACL) 2022 • Duy-Hung Nguyen, Nguyen Viet Dung Nghiem, Bao-Sinh Nguyen, Dung Tien Le, Shahab Sabahi, Minh-Tien Nguyen, Hung Le
For summarization, human preference is critical to tame outputs of the summarizer in favor of human interests, as ground-truth summaries are scarce and ambiguous.
no code implementations • 24 Feb 2022 • Kien Do, Haripriya Harikumar, Hung Le, Dung Nguyen, Truyen Tran, Santu Rana, Dang Nguyen, Willy Susilo, Svetha Venkatesh
Trojan attacks on deep neural networks are both dangerous and surreptitious.
1 code implementation • 3 Dec 2021 • Hung Le, Majid Abdolshah, Thommen K. George, Kien Do, Dung Nguyen, Svetha Venkatesh
We introduce a novel training procedure for policy gradient methods wherein episodic memory is used to optimize the hyperparameters of reinforcement learning algorithms on-the-fly.
no code implementations • 13 Nov 2021 • Duy-Hung Nguyen, Bao-Sinh Nguyen, Nguyen Viet Dung Nghiem, Dung Tien Le, Mim Amina Khatun, Minh-Tien Nguyen, Hung Le
Automatic summarization of legal texts is an important and still a challenging task since legal documents are often long and complicated with unusual structures and styles.
no code implementations • 3 Nov 2021 • Thommen George Karimpanal, Hung Le, Majid Abdolshah, Santu Rana, Sunil Gupta, Truyen Tran, Svetha Venkatesh
The optimistic nature of the Q-learning target leads to an overestimation bias, which is an inherent problem associated with standard $Q-$learning.
no code implementations • NeurIPS 2021 • Hung Le, Thommen Karimpanal George, Majid Abdolshah, Truyen Tran, Svetha Venkatesh
Episodic control enables sample efficiency in reinforcement learning by recalling past experiences from an episodic memory.
no code implementations • ICLR 2022 • Kha Pham, Hung Le, Man Ngo, Truyen Tran, Bao Ho, Svetha Venkatesh
We propose Generative Pseudo-Inverse Memory (GPM), a class of deep generative memory models that are fast to write in and read out.
no code implementations • 29 Sep 2021 • Thommen Karimpanal George, Majid Abdolshah, Hung Le, Santu Rana, Sunil Gupta, Truyen Tran, Svetha Venkatesh
The objective in goal-based reinforcement learning is to learn a policy to reach a particular goal state within the environment.
no code implementations • 29 Sep 2021 • Majid Abdolshah, Hung Le, Thommen Karimpanal George, Vuong Le, Sunil Gupta, Santu Rana, Svetha Venkatesh
Whilst Generative Adversarial Networks (GANs) generate visually appealing high resolution images, the latent representations (or codes) of these models do not allow controllable changes on the semantic attributes of the generated images.
no code implementations • 20 Aug 2021 • Majid Abdolshah, Hung Le, Thommen Karimpanal George, Sunil Gupta, Santu Rana, Svetha Venkatesh
This is achieved by representing the global transition dynamics as a union of local transition functions, each with respect to one active object in the scene.
no code implementations • 18 Jul 2021 • Majid Abdolshah, Hung Le, Thommen Karimpanal George, Sunil Gupta, Santu Rana, Svetha Venkatesh
Transfer in reinforcement learning is usually achieved through generalisation across tasks.
1 code implementation • 3 Jul 2021 • Hung Le
Artificial neural networks model neurons and synapses in the brain by interconnecting computational units via weights, which is a typical class of machine learning algorithms that resembles memory structure.
no code implementations • 16 Jun 2021 • Hung Le, Nancy F. Chen, Steven C. H. Hoi
Video-grounded dialogue systems aim to integrate video understanding and dialogue understanding to generate responses that are relevant to both the dialogue and video context.
no code implementations • 16 Apr 2021 • Hung Le, Nancy F. Chen, Steven C. H. Hoi
Neural module networks (NMN) have achieved success in image-grounded tasks such as Visual Question Answering (VQA) on synthetic images.
no code implementations • ICLR 2021 • Hung Le, Nancy F. Chen, Steven C. H. Hoi
PDC model then learns to predict reasoning paths over this semantic graph.
1 code implementation • ACL 2021 • Hung Le, Chinnadhurai Sankar, Seungwhan Moon, Ahmad Beirami, Alborz Geramifard, Satwik Kottur
A video-grounded dialogue system is required to understand both dialogue, which contains semantic dependencies from turn to turn, and video, which contains visual cues of spatial and temporal scene variations.
no code implementations • 1 Jan 2021 • Hung Le, Nancy F. Chen, Steven Hoi
Neural module networks (NMN) have achieved success in image-grounded tasks such as question answering (QA) on synthetic images.
1 code implementation • EMNLP 2020 • Hung Le, Doyen Sahoo, Nancy F. Chen, Steven C. H. Hoi
Video-grounded dialogues are very challenging due to (i) the complexity of videos which contain both spatial and temporal variations, and (ii) the complexity of user utterances which query different segments and/or different objects in videos over multiple dialogue turns.
no code implementations • NeurIPS 2021 • Hung Le, Svetha Venkatesh
For the first time a Neural Program is treated as a datum in memory, paving the ways for modular, recursive and procedural neural programming.
no code implementations • ACL 2020 • Hung Le, Steven C. H. Hoi
Pre-trained language models have shown remarkable success in improving various downstream NLP tasks due to their ability to capture dependencies in textual data and generate natural responses.
1 code implementation • EMNLP 2020 • Hung Le, Doyen Sahoo, Chenghao Liu, Nancy F. Chen, Steven C. H. Hoi
Building an end-to-end conversational agent for multi-domain task-oriented dialogues has been an open challenge for two main reasons.
no code implementations • 25 Feb 2020 • Hung Le, Nancy F. Chen
Audio-Visual Scene-Aware Dialog (AVSD) is an extension from Video Question Answering (QA) whereby the dialogue agent is required to generate natural language responses to address user queries and carry on conversations.
1 code implementation • ICLR 2020 • Hung Le, Richard Socher, Steven C. H. Hoi
Recent efforts in Dialogue State Tracking (DST) for task-oriented dialogues have progressed toward open-vocabulary or generation-based approaches where the models can generate slot value candidates from the dialogue history itself.
Ranked #13 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.0
1 code implementation • ICML 2020 • Hung Le, Truyen Tran, Svetha Venkatesh
Heretofore, neural networks with external memory are restricted to single memory with lossy representations of memory interactions.
Ranked #1 on Question Answering on bAbi
no code implementations • 5 Nov 2019 • Duc Nguyen, Nhan Tran, Hung Le
Convolutional Recurrent Neural Networks (CRNNs) excel at scene text recognition.
Optical Character Recognition Optical Character Recognition (OCR) +1
no code implementations • 26 Sep 2019 • Doyen Sahoo, Wang Hao, Shu Ke, Wu Xiongwei, Hung Le, Palakorn Achananuparp, Ee-Peng Lim, Steven C. H. Hoi
FoodAI has made food logging convenient, aiding smart consumption and a healthy lifestyle.
1 code implementation • ACL 2019 • Hung Le, Doyen Sahoo, Nancy F. Chen, Steven C. H. Hoi
Developing Video-Grounded Dialogue Systems (VGDS), where a dialogue is conducted based on visual and audio aspects of a given video, is significantly more challenging than traditional image or text-grounded dialogue systems because (1) feature space of videos span across multiple picture frames, making it difficult to obtain semantic information; and (2) a dialogue agent must perceive and process information from different modalities (audio, video, caption, etc.)
Ranked #4 on Response Generation on SIMMC2.0
1 code implementation • ICLR 2020 • Hung Le, Truyen Tran, Svetha Venkatesh
Neural networks powered with external memory simulate computer behaviors.
Ranked #5 on Question Answering on bAbi (Mean Error Rate metric)
no code implementations • ICLR 2019 • Doyen Sahoo, Hung Le, Chenghao Liu, Steven C. H. Hoi
Most existing work assumes that both training and test tasks are drawn from the same distribution, and a large amount of labeled data is available in the training tasks.
1 code implementation • ICLR 2019 • Hung Le, Truyen Tran, Svetha Venkatesh
Memory-augmented neural networks consisting of a neural controller and an external memory have shown potentials in long-term sequential learning.
Ranked #5 on Text Classification on Yahoo! Answers
1 code implementation • NeurIPS 2018 • Hung Le, Truyen Tran, Thin Nguyen, Svetha Venkatesh
Introducing variability while maintaining coherence is a core task in learning to generate utterances in conversation.
no code implementations • 11 Feb 2018 • Hung Le, Truyen Tran, Svetha Venkatesh
The decoding controller generates a treatment sequence, one treatment option at a time.
3 code implementations • 9 Feb 2018 • Hung Le, Quang Pham, Doyen Sahoo, Steven C. H. Hoi
This approach allows the model to capture several types of semantic information, which was not possible by the existing models.
1 code implementation • 3 Feb 2018 • Asjad Khan, Hung Le, Kien Do, Truyen Tran, Aditya Ghose, Hoa Dam, Renuka Sindhgatta
Process-aware Recommender systems can provide critical decision support functionality to aid business process execution by recommending what actions to take next.
1 code implementation • 2 Feb 2018 • Hung Le, Truyen Tran, Svetha Venkatesh
One of the core tasks in multi-view learning is to capture relations among views.
no code implementations • 19 May 2017 • Hung Le, Ali Borji
In this work, we explain in detail how receptive fields, effective receptive fields, and projective fields of neurons in different layers, convolution or pooling, of a Convolutional Neural Network (CNN) are calculated.