no code implementations • Findings (EMNLP) 2021 • Fahimeh Saleh, Wray Buntine, Gholamreza Haffari, Lan Du
Multilingual Neural Machine Translation (MNMT) trains a single NMT model that supports translation between multiple languages, rather than training separate models for different languages.
no code implementations • 26 Nov 2024 • Yuanyuan Qi, Jueqing Lu, Xiaohao Yang, Joanne Enticott, Lan Du
The primary challenge of multi-label active learning, differing it from multi-class active learning, lies in assessing the informativeness of an indefinite number of labels while also accounting for the inherited label correlation.
no code implementations • 13 Nov 2024 • Xiaohao Yang, He Zhao, Weijie Xu, Yuanyuan Qi, Jueqing Lu, Dinh Phung, Lan Du
While Large Language Models (LLMs) have demonstrated promising capabilities in topic discovery, their direct application to topic modeling suffers from issues such as incomplete topic coverage, misalignment of topics, and inefficiency.
1 code implementation • 22 Jul 2024 • Zeyu Wang, Jingyu Lin, Yifei Qian, Yi Huang, Shicen Tian, Bosong Chai, Juncan Deng, Qu Yang, Lan Du, Cunjian Chen, Kejie Huang
However, most diffusion models are limited to visible RGB image generation.
1 code implementation • 13 Jun 2024 • Xiaohao Yang, He Zhao, Dinh Phung, Wray Buntine, Lan Du
Topic modeling has been a widely used tool for unsupervised text analysis.
1 code implementation • 7 Jun 2024 • Yixin Huang, Yiqi Jin, Ke Tao, Kaijian Xia, Jianfeng Gu, Lei Yu, Lan Du, Cunjian Chen
In this paper, we present a 3D-based deep learning approach called MTS-Net for diagnosing May-Thurner Syndrome using CT scans.
no code implementations • 3 Jun 2024 • Jueqing Lu, Lan Du, Wray Buntine, Myong Chol Jung, Joanna Dipnall, Belinda Gabbe
Resolving conflicts is essential to make the decisions of multi-view classification more reliable.
no code implementations • 25 May 2024 • Myong Chol Jung, He Zhao, Joanna Dipnall, Belinda Gabbe, Lan Du
In this study, we investigate the near OOD detection capabilities of prompt learning models and observe that commonly used OOD scores have limited performance in near OOD detection.
no code implementations • 2 Apr 2024 • Lin Li, Jianping Gou, Baosheng Yu, Lan Du, Zhang Yiand Dacheng Tao
Federated Learning (FL) seeks to train a model collaboratively without sharing private training data from individual clients.
no code implementations • 15 Jan 2024 • Wei Tan, Ngoc Dang Nguyen, Lan Du, Wray Buntine
Within the scope of natural language processing, the domain of multi-label text classification is uniquely challenging due to its expansive and uneven label distribution.
no code implementations • 15 Dec 2023 • Wei Tan, Lan Du, Wray Buntine
Expected Loss Reduction (ELR) focuses on a Bayesian estimate of the reduction in classification error, and more general costs fit in the same framework.
no code implementations • 2 Nov 2023 • Haocheng Luo, Wei Tan, Ngoc Dang Nguyen, Lan Du
Active learning, a widely adopted technique for enhancing machine learning models in text and image classification tasks with limited annotation resources, has received relatively little attention in the domain of Named Entity Recognition (NER).
no code implementations • 2 Nov 2023 • Ngoc Dang Nguyen, Wei Tan, Lan Du, Wray Buntine, Richard Beare, Changyou Chen
Named entity recognition (NER), a task that identifies and categorizes named entities such as persons or organizations from text, is traditionally framed as a multi-class classification problem.
1 code implementation • 24 Jul 2023 • Xiaohao Yang, He Zhao, Dinh Phung, Lan Du
To do so, we propose to enhance NTMs by narrowing the semantic distance between similar documents, with the underlying assumption that documents from different corpora may share similar semantics.
no code implementations • 15 Apr 2023 • Jionghao Lin, Wei Tan, Ngoc Dang Nguyen, David Lang, Lan Du, Wray Buntine, Richard Beare, Guanliang Chen, Dragan Gasevic
We note that many prior studies on classifying educational DAs employ cross entropy (CE) loss to optimize DA classifiers on low-resource data with imbalanced DA distribution.
no code implementations • 12 Apr 2023 • Wei Tan, Jionghao Lin, David Lang, Guanliang Chen, Dragan Gasevic, Lan Du, Wray Buntine
Then, the study investigates how the AL methods can select informative samples to support DA classifiers in the AL sampling process.
1 code implementation • 27 Feb 2023 • Weidong Chen, Xiaofen Xing, Xiangmin Xu, Jianxin Pang, Lan Du
Paralinguistic speech processing is important in addressing many issues, such as sentiment and neurocognitive disorder analyses.
1 code implementation • 23 Dec 2022 • Fucai Ke, Weiqing Wang, Weicong Tan, Lan Du, Yuan Jin, Yujin Huang, Hongzhi Yin
Knowledge tracing (KT) aims to leverage students' learning histories to estimate their mastery levels on a set of pre-defined skills, based on which the corresponding future performance can be accurately predicted.
no code implementations • 9 Dec 2022 • Ngoc Dang Nguyen, Wei Tan, Wray Buntine, Richard Beare, Changyou Chen, Lan Du
To the best of our knowledge, this is the first work that brings AUC maximization to the NER setting.
Low Resource Named Entity Recognition
named-entity-recognition
+2
no code implementations • 11 Nov 2022 • Ngoc Dang Nguyen, Lan Du, Wray Buntine, Changyou Chen, Richard Beare
Domain adaptation is an effective solution to data scarcity in low-resource scenarios.
1 code implementation • 7 Nov 2022 • Erxin Yu, Lan Du, Yuan Jin, Zhepei Wei, Yi Chang
Recently, discrete latent variable models have received a surge of interest in both Natural Language Processing (NLP) and Computer Vision (CV), attributed to their comparable performance to the continuous counterparts in representation learning, while being more interpretable in their predictions.
no code implementations • 6 Oct 2022 • Myong Chol Jung, He Zhao, Joanna Dipnall, Belinda Gabbe, Lan Du
Uncertainty estimation is essential to make neural networks trustworthy in real-world applications.
1 code implementation • NeurIPS 2021 • Wei Tan, Lan Du, Wray Buntine
We convert the ELR framework to estimate the increase in (strictly proper) scores like log probability or negative mean square error, which we call Bayesian Estimate of Mean Proper Scores (BEMPS).
no code implementations • 15 Oct 2021 • Fahimeh Saleh, Wray Buntine, Gholamreza Haffari, Lan Du
Multilingual Neural Machine Translation (MNMT) trains a single NMT model that supports translation between multiple languages, rather than training separate models for different languages.
no code implementations • Findings (EMNLP) 2021 • Kelvin Lo, Yuan Jin, Weicong Tan, Ming Liu, Lan Du, Wray Buntine
This paper proposes a transformer over transformer framework, called Transformer$^2$, to perform neural text segmentation.
no code implementations • EMNLP 2021 • Yuan Jin, He Zhao, Ming Liu, Lan Du, Wray Buntine
Neural topic models (NTMs) apply deep neural networks to topic modelling.
no code implementations • Findings (EMNLP) 2021 • Jiaxin Ju, Ming Liu, Huan Yee Koh, Yuan Jin, Lan Du, Shirui Pan
This paper presents an unsupervised extractive approach to summarize scientific long documents based on the Information Bottleneck principle.
no code implementations • 28 Aug 2021 • Stella Ho, Ming Liu, Lan Du, Longxiang Gao, Yong Xiang
Continual learning (CL) refers to a machine learning paradigm that learns continuously without forgetting previously acquired knowledge.
no code implementations • 8 Mar 2021 • Man Wu, Shirui Pan, Lan Du, Xingquan Zhu
By generating multiple graphs at different distance levels, based on the adjacency matrix, we develop a long-short distance attention model to model these graphs.
2 code implementations • 5 Mar 2021 • Maximillian Merrillees, Lan Du
Extreme multi-label classification (XML) is becoming increasingly relevant in the era of big data.
Extreme Multi-Label Classification
MUlTI-LABEL-ClASSIFICATION
no code implementations • 28 Feb 2021 • He Zhao, Dinh Phung, Viet Huynh, Yuan Jin, Lan Du, Wray Buntine
Topic modelling has been a successful technique for text analysis for almost twenty years.
no code implementations • 21 Jan 2021 • Liyuan Sun, Jianping Gou, Baosheng Yu, Lan Du, DaCheng Tao
However, most of the existing knowledge distillation methods consider only one type of knowledge learned from either instance features or instance relations via a specific distillation strategy in teacher-student learning.
1 code implementation • EMNLP 2020 • Jueqing Lu, Lan Du, Ming Liu, Joanna Dipnall
Few/Zero-shot learning is a big challenge of many classifications tasks, where a classifier is required to recognise instances of classes that have very few or even no training samples.
1 code implementation • 17 Jul 2020 • Jinming Zhao, Ming Liu, Longxiang Gao, Yuan Jin, Lan Du, He Zhao, He Zhang, Gholamreza Haffari
Obtaining training data for multi-document summarization (MDS) is time consuming and resource-intensive, so recent neural models can only be trained for limited domains.
no code implementations • 21 Feb 2020 • Yuan Jin, He Zhao, Ming Liu, Ye Zhu, Lan Du, Longxiang Gao, He Zhang, Yunfeng Li
Based on the ELBOs, we propose a VAE-based Bayesian MF framework.
no code implementations • 12 Oct 2019 • Yuan Jin, Ming Liu, Yunfeng Li, Ruohua Xu, Lan Du, Longxiang Gao, Yong Xiang
Under synthetic data evaluation, VAE-BPTF tended to recover the right number of latent factors and posterior parameter values.
no code implementations • ACL 2019 • He Zhao, Lan Du, Guanfeng Liu, Wray Buntine
Short texts such as tweets often contain insufficient word co-occurrence information for training conventional topic models.
1 code implementation • 2 May 2019 • He Zhao, Piyush Rai, Lan Du, Wray Buntine, Mingyuan Zhou
Many applications, such as text modelling, high-throughput sequencing, and recommender systems, require analysing sparse, high-dimensional, and overdispersed discrete (count-valued or binary) data.
2 code implementations • NeurIPS 2018 • He Zhao, Lan Du, Wray Buntine, Mingyuan Zhou
Recently, considerable research effort has been devoted to developing deep architectures for topic models to learn topic structures.
no code implementations • TACL 2015 • Dat Quoc Nguyen, Richard Billingsley, Lan Du, Mark Johnson
Probabilistic topic models are widely used to discover latent topics in document collections, while latent feature vector representations of words have been used to obtain high performance in many NLP tasks.
1 code implementation • ICML 2018 • He Zhao, Lan Du, Wray Buntine, Mingyuan Zhou
One important task of topic modeling for text analysis is interpretability.
1 code implementation • 19 Sep 2017 • He Zhao, Lan Du, Wray Buntine, Gang Liu
Besides the text content, documents and their associated words usually come with rich sets of meta informa- tion, such as categories of documents and semantic/syntactic features of words, like those encoded in word embeddings.
no code implementations • ACL 2017 • Shervin Malmasi, Mark Dras, Mark Johnson, Lan Du, Magdalena Wolska
Most work on segmenting text does so on the basis of topic changes, but it can be of interest to segment by other, stylistically expressed characteristics such as change of authorship or native language.
1 code implementation • ICML 2017 • He Zhao, Lan Du, Wray Buntine
Relational data are usually highly incomplete in practice, which inspires us to leverage side information to improve the performance of community detection and link prediction.
no code implementations • 22 Sep 2016 • Kar Wai Lim, Wray Buntine, Changyou Chen, Lan Du
In this article, we present efficient methods for the use of these processes in this hierarchical context, and apply them to latent variable models for text analytics.
no code implementations • NeurIPS 2009 • Lan Du, Lu Ren, Lawrence Carin, David B. Dunson
The model clusters the images into classes, and each image is segmented into a set of objects, also allowing the opportunity to assign a word to each object (localized labeling).