Search Results for author: Wray Buntine

Found 59 papers, 20 papers with code

All Labels Are Not Created Equal: Enhancing Semi-supervision via Label Grouping and Co-training

1 code implementation CVPR 2021 Islam Nassar, Samitha Herath, Ehsan Abbasnejad, Wray Buntine, Gholamreza Haffari

We train two classifiers with two different views of the class labels: one classifier uses the one-hot view of the labels and disregards any potential similarity among the classes, while the other uses a distributed view of the labels and groups potentially similar classes together.

Semi-Supervised Image Classification

Cross-Domain Graph Anomaly Detection via Anomaly-aware Contrastive Alignment

1 code implementation2 Dec 2022 Qizhou Wang, Guansong Pang, Mahsa Salehi, Wray Buntine, Christopher Leckie

In doing so, ACT effectively transfers anomaly-informed knowledge from the source graph to learn the complex node relations of the normal class for GAD on the target graph without any specification of the anomaly distributions.

Contrastive Learning Domain Adaptation +1

PiVe: Prompting with Iterative Verification Improving Graph-based Generative Capability of LLMs

1 code implementation21 May 2023 Jiuzhou Han, Nigel Collier, Wray Buntine, Ehsan Shareghi

We show how a small language model could be trained to act as a verifier module for the output of an LLM(i. e., ChatGPT, GPT-4), and to iteratively improve its performance via fine-grained corrective instructions.

Data Augmentation Graph Generation +1

$\mathtt{MedGraph:}$ Structural and Temporal Representation Learning of Electronic Medical Records

1 code implementation8 Dec 2019 Bhagya Hettige, Yuan-Fang Li, Weiqing Wang, Suong Le, Wray Buntine

To address these limitations, we present $\mathtt{MedGraph}$, a supervised EMR embedding method that captures two types of information: (1) the visit-code associations in an attributed bipartite graph, and (2) the temporal sequencing of visits through a point process.

Attribute Point Processes +1

Variational Autoencoders for Sparse and Overdispersed Discrete Data

1 code implementation2 May 2019 He Zhao, Piyush Rai, Lan Du, Wray Buntine, Mingyuan Zhou

Many applications, such as text modelling, high-throughput sequencing, and recommender systems, require analysing sparse, high-dimensional, and overdispersed discrete (count-valued or binary) data.

Collaborative Filtering Multi-Label Learning +1

Neural Topic Model via Optimal Transport

1 code implementation ICLR 2021 He Zhao, Dinh Phung, Viet Huynh, Trung Le, Wray Buntine

Recently, Neural Topic Models (NTMs) inspired by variational autoencoders have obtained increasingly research interest due to their promising results on text analysis.

Topic Models

Learning How to Actively Learn: A Deep Imitation Learning Approach

1 code implementation ACL 2018 Ming Liu, Wray Buntine, Gholamreza Haffari

Heuristic-based active learning (AL) methods are limited when the data distribution of the underlying learning problems vary.

Active Learning General Classification +8

Gaussian Embedding of Large-scale Attributed Graphs

1 code implementation2 Dec 2019 Bhagya Hettige, Yuan-Fang Li, Weiqing Wang, Wray Buntine

Graph embedding methods transform high-dimensional and complex graph contents into low-dimensional representations.

Graph Embedding Link Prediction +1

Accurate parameter estimation for Bayesian Network Classifiers using Hierarchical Dirichlet Processes

4 code implementations25 Aug 2017 Francois Petitjean, Wray Buntine, Geoffrey I. Webb, Nayyar Zaidi

The main result of this paper is to show that improved parameter estimation allows BNCs to outperform leading learning methods such as Random Forest for both 0-1 loss and RMSE, albeit just on categorical datasets.

General Classification

Diversity Enhanced Active Learning with Strictly Proper Scoring Rules

1 code implementation NeurIPS 2021 Wei Tan, Lan Du, Wray Buntine

We convert the ELR framework to estimate the increase in (strictly proper) scores like log probability or negative mean square error, which we call Bayesian Estimate of Mean Proper Scores (BEMPS).

Active Learning text-classification +1

MetaLDA: a Topic Model that Efficiently Incorporates Meta information

1 code implementation19 Sep 2017 He Zhao, Lan Du, Wray Buntine, Gang Liu

Besides the text content, documents and their associated words usually come with rich sets of meta informa- tion, such as categories of documents and semantic/syntactic features of words, like those encoded in word embeddings.

Topic Models Word Embeddings

Dirichlet belief networks for topic structure learning

2 code implementations NeurIPS 2018 He Zhao, Lan Du, Wray Buntine, Mingyuan Zhou

Recently, considerable research effort has been devoted to developing deep architectures for topic models to learn topic structures.

Topic Models

The Neglected Sibling: Isotropic Gaussian Posterior for VAE

1 code implementation14 Oct 2021 Lan Zhang, Wray Buntine, Ehsan Shareghi

Deep generative models have been widely used in several areas of NLP, and various techniques have been proposed to augment them or address their training challenges.

On the Effect of Isotropy on VAE Representations of Text

1 code implementation ACL 2022 Lan Zhang, Wray Buntine, Ehsan Shareghi

Injecting desired geometric properties into text representations has attracted a lot of attention.

Reward Engineering for Generating Semi-structured Explanation

1 code implementation15 Sep 2023 Jiuzhou Han, Wray Buntine, Ehsan Shareghi

Semi-structured explanation depicts the implicit process of a reasoner with an explicit representation.

Explanation Generation Reinforcement Learning (RL)

Leveraging Node Attributes for Incomplete Relational Data

1 code implementation ICML 2017 He Zhao, Lan Du, Wray Buntine

Relational data are usually highly incomplete in practice, which inspires us to leverage side information to improve the performance of community detection and link prediction.

Community Detection Link Prediction

Unsupervised Sentence Simplification via Dependency Parsing

1 code implementation10 Jun 2022 Vy Vo, Weiqing Wang, Wray Buntine

Text simplification is the task of rewriting a text so that it is readable and easily understood.

Dependency Parsing Sentence +2

SQAPlanner: Generating Data-Informed Software Quality Improvement Plans

1 code implementation19 Feb 2021 Dilini Rajapaksha, Chakkrit Tantithamthavorn, Jirayus Jiarpakdee, Christoph Bergmeir, John Grundy, Wray Buntine

Thus, our SQAPlanner paves a way for novel research in actionable software analytics-i. e., generating actionable guidance on what should practitioners do and not do to decrease the risk of having defects to support SQA planning.

Distinguishing Question Subjectivity from Difficulty for Improved Crowdsourcing

no code implementations12 Feb 2018 Yuan Jin, Mark Carman, Ye Zhu, Wray Buntine

Experiments show that our model(1) improves the performance of both quality control for crowd-sourced answers and next answer prediction for crowd-workers, and (2) can potentially provide coherent rankings of questions in terms of their difficulty and subjectivity, so that task providers can refine their designs of the crowdsourcing tasks, e. g. by removing highly subjective questions or inappropriately difficult questions.

Clustering

Bibliographic Analysis with the Citation Network Topic Model

no code implementations22 Sep 2016 Kar Wai Lim, Wray Buntine

Bibliographic analysis considers author's research areas, the citation network and paper content among other things.

Clustering

Twitter-Network Topic Model: A Full Bayesian Treatment for Social Network and Text Modeling

no code implementations22 Sep 2016 Kar Wai Lim, Changyou Chen, Wray Buntine

Exploiting this additional information, we propose the Twitter-Network (TN) topic model to jointly model the text and the social network in a full Bayesian nonparametric way.

Topic Models

Nonparametric Bayesian Topic Modelling with the Hierarchical Pitman-Yor Processes

no code implementations22 Sep 2016 Kar Wai Lim, Wray Buntine, Changyou Chen, Lan Du

In this article, we present efficient methods for the use of these processes in this hierarchical context, and apply them to latent variable models for text analytics.

Topic Models

Twitter Opinion Topic Model: Extracting Product Opinions from Tweets by Leveraging Hashtags and Sentiment Lexicon

no code implementations21 Sep 2016 Kar Wai Lim, Wray Buntine

Although social media data like tweets are laden with opinions, their "dirty" nature (as natural language) has discouraged researchers from applying LDA-based opinion model for product review mining.

Opinion Mining Sentiment Analysis

Bibliographic Analysis on Research Publications using Authors, Categorical Labels and the Citation Network

no code implementations21 Sep 2016 Kar Wai Lim, Wray Buntine

In this paper, we combine these three in a topic model that produces a bibliographic model of authors, topics and documents, using a nonparametric extension of a combination of the Poisson mixed-topic link model and the author-topic model.

Clustering

Word Features for Latent Dirichlet Allocation

no code implementations NeurIPS 2010 James Petterson, Wray Buntine, Shravan M. Narayanamurthy, Tibério S. Caetano, Alex J. Smola

We extend Latent Dirichlet Allocation (LDA) by explicitly allowing for the encoding of side information in the distribution over words.

Leveraging Meta Information in Short Text Aggregation

no code implementations ACL 2019 He Zhao, Lan Du, Guanfeng Liu, Wray Buntine

Short texts such as tweets often contain insufficient word co-occurrence information for training conventional topic models.

Clustering Topic Models

LoRMIkA: Local rule-based model interpretability with k-optimal associations

no code implementations11 Aug 2019 Dilini Rajapaksha, Christoph Bergmeir, Wray Buntine

In this paper, we propose Local Rule-based Model Interpretability with k-optimal Associations (LoRMIkA), a novel model-agnostic approach that obtains k-optimal association rules from a neighbourhood of the instance to be explained.

BIG-bench Machine Learning counterfactual +1

Discriminative, Generative and Self-Supervised Approaches for Target-Agnostic Learning

no code implementations12 Nov 2020 Yuan Jin, Wray Buntine, Francois Petitjean, Geoffrey I. Webb

For this task, we survey a wide range of techniques available for handling missing values, self-supervised training and pseudo-likelihood training, and adapt them to a suite of algorithms that are suitable for the task.

Self-Supervised Learning

Temporal Cascade and Structural Modelling of EHRs for Granular Readmission Prediction

no code implementations4 Feb 2021 Bhagya Hettige, Weiqing Wang, Yuan-Fang Li, Suong Le, Wray Buntine

Although a point process (e. g., Hawkes process) is able to model a cascade temporal relationship, it strongly relies on a prior generative process assumption.

Decision Making Point Processes +1

Topic Modelling Meets Deep Neural Networks: A Survey

no code implementations28 Feb 2021 He Zhao, Dinh Phung, Viet Huynh, Yuan Jin, Lan Du, Wray Buntine

Topic modelling has been a successful technique for text analysis for almost twenty years.

Navigate Text Generation +1

Topic Model or Topic Twaddle? Re-evaluating Semantic Interpretability Measures

no code implementations NAACL 2021 Caitlin Doogan, Wray Buntine

In this paper, we probe the issue of validity in topic model evaluation and assess how informative coherence measures are for specialized collections used in an applied setting.

Topic Models

Multilingual Neural Machine Translation:Can Linguistic Hierarchies Help?

no code implementations15 Oct 2021 Fahimeh Saleh, Wray Buntine, Gholamreza Haffari, Lan Du

Multilingual Neural Machine Translation (MNMT) trains a single NMT model that supports translation between multiple languages, rather than training separate models for different languages.

Knowledge Distillation Machine Translation +2

Multilingual Neural Machine Translation: Can Linguistic Hierarchies Help?

no code implementations Findings (EMNLP) 2021 Fahimeh Saleh, Wray Buntine, Gholamreza Haffari, Lan Du

Multilingual Neural Machine Translation (MNMT) trains a single NMT model that supports translation between multiple languages, rather than training separate models for different languages.

Knowledge Distillation Machine Translation +2

Robust Educational Dialogue Act Classifiers with Low-Resource and Imbalanced Datasets

no code implementations15 Apr 2023 Jionghao Lin, Wei Tan, Ngoc Dang Nguyen, David Lang, Lan Du, Wray Buntine, Richard Beare, Guanliang Chen, Dragan Gasevic

We note that many prior studies on classifying educational DAs employ cross entropy (CE) loss to optimize DA classifiers on low-resource data with imbalanced DA distribution.

A Survey on Out-of-Distribution Evaluation of Neural NLP Models

no code implementations27 Jun 2023 Xinzhe Li, Ming Liu, Shang Gao, Wray Buntine

Adversarial robustness, domain generalization and dataset biases are three active lines of research contributing to out-of-distribution (OOD) evaluation on neural NLP models.

Adversarial Robustness Domain Generalization

Low-Resource Named Entity Recognition: Can One-vs-All AUC Maximization Help?

no code implementations2 Nov 2023 Ngoc Dang Nguyen, Wei Tan, Lan Du, Wray Buntine, Richard Beare, Changyou Chen

Named entity recognition (NER), a task that identifies and categorizes named entities such as persons or organizations from text, is traditionally framed as a multi-class classification problem.

Low Resource Named Entity Recognition Meta-Learning +4

HOMOE: A Memory-Based and Composition-Aware Framework for Zero-Shot Learning with Hopfield Network and Soft Mixture of Experts

no code implementations23 Nov 2023 Do Huu Dat, Po Yuan Mao, Tien Hoang Nguyen, Wray Buntine, Mohammed Bennamoun

In our paper, we propose a novel framework that for the first time combines the Modern Hopfield Network with a Mixture of Experts (HOMOE) to classify the compositions of previously unseen objects.

Compositional Zero-Shot Learning

Bayesian Estimate of Mean Proper Scores for Diversity-Enhanced Active Learning

no code implementations15 Dec 2023 Wei Tan, Lan Du, Wray Buntine

Expected Loss Reduction (ELR) focuses on a Bayesian estimate of the reduction in classification error, and more general costs fit in the same framework.

Active Learning

Harnessing the Power of Beta Scoring in Deep Active Learning for Multi-Label Text Classification

no code implementations15 Jan 2024 Wei Tan, Ngoc Dang Nguyen, Lan Du, Wray Buntine

Within the scope of natural language processing, the domain of multi-label text classification is uniquely challenging due to its expansive and uneven label distribution.

Active Learning Multi Label Text Classification +2

Towards Uncertainty-Aware Language Agent

no code implementations25 Jan 2024 Jiuzhou Han, Wray Buntine, Ehsan Shareghi

We present the Uncertainty-Aware Language Agent (UALA), a framework that orchestrates the interaction between the agent and the external world using uncertainty quantification.

StrategyQA Uncertainty Quantification

OntoMedRec: Logically-Pretrained Model-Agnostic Ontology Encoders for Medication Recommendation

no code implementations29 Jan 2024 Weicong Tan, Weiqing Wang, Xin Zhou, Wray Buntine, Gordon Bingham, Hongzhi Yin

Most existing medication recommendation models learn representations for medical concepts based on electronic health records (EHRs) and make recommendations with learnt representations.

Improving Vietnamese-English Medical Machine Translation

no code implementations28 Mar 2024 Nhu Vo, Dat Quoc Nguyen, Dung D. Le, Massimo Piccardi, Wray Buntine

Machine translation for Vietnamese-English in the medical domain is still an under-explored research area.

Machine Translation Sentence +1

Cannot find the paper you are looking for? You can Submit a new open access paper.