no code implementations • EMNLP (FEVER) 2021 • Mohammed Saeed, Giulio Alfarano, Khai Nguyen, Duc Pham, Raphael Troncy, Paolo Papotti
Computational fact-checking has gained a lot of traction in the machine learning and natural language processing communities.
no code implementations • 8 Feb 2025 • Manh Luong, Khai Nguyen, Dinh Phung, Gholamreza Haffari, Lizhen Qu
However, the contrastive method ignores the temporal information when measuring similarity across acoustic and linguistic modalities, leading to inferior performance.
no code implementations • 31 Jan 2025 • Khai Nguyen, Hai Nguyen, Tuan Pham, Nhat Ho
We introduce sliced optimal transport dataset distance (s-OTDD), a model-agnostic, embedding-agnostic approach for dataset comparison that requires no training, is robust to variations in the number of classes, and can handle disjoint label sets.
no code implementations • 22 Nov 2024 • Khai Nguyen, Peter Mueller
The second, sliced mixture Wasserstein, leverages the linearity of Gaussian mixture measures for efficient projection.
no code implementations • 21 May 2024 • Nicola Bariletto, Khai Nguyen, Nhat Ho
This paper presents a novel optimization framework to address key challenges presented by modern machine learning applications: High dimensionality, distributional uncertainty, and data heterogeneity.
1 code implementation • 16 May 2024 • Manh Luong, Khai Nguyen, Nhat Ho, Reza Haf, Dinh Phung, Lizhen Qu
The Learning-to-match (LTM) framework proves to be an effective inverse optimal transport approach for learning the underlying ground metric between two sources of data, facilitating subsequent matching.
no code implementations • 13 May 2024 • Khai Nguyen, Hai Nguyen, Nhat Ho
The sliced Wasserstein barycenter (SWB) is a widely acknowledged method for efficiently generalizing the averaging operation within probability measure spaces.
1 code implementation • 23 Apr 2024 • Khai Nguyen, Nhat Ho
By using HHRT, we extend the SW into Hierarchical Hybrid Sliced Wasserstein (H2SW) distance which is designed specifically for comparing heterogeneous joint distributions.
no code implementations • CVPR 2024 • Tung Le, Khai Nguyen, Shanlin Sun, Nhat Ho, Xiaohui Xie
In the realm of computer vision and graphics, accurately establishing correspondences between geometric 3D shapes is pivotal for applications like object tracking, registration, texture transfer, and statistical shape analysis.
no code implementations • 7 Feb 2024 • Huy Nguyen, Khai Nguyen, Nhat Ho
We consider the parameter estimation problem in the deviated Gaussian mixture of experts in which the data are generated from $(1 - \lambda^{\ast}) g_0(Y| X)+ \lambda^{\ast} \sum_{i = 1}^{k_{\ast}} p_{i}^{\ast} f(Y|(a_{i}^{\ast})^{\top}X+b_i^{\ast},\sigma_{i}^{\ast})$, where $X, Y$ are respectively a covariate vector and a response variable, $g_{0}(Y|X)$ is a known function, $\lambda^{\ast} \in [0, 1]$ is true but unknown mixing proportion, and $(p_{i}^{\ast}, a_{i}^{\ast}, b_{i}^{\ast}, \sigma_{i}^{\ast})$ for $1 \leq i \leq k^{\ast}$ are unknown parameters of the Gaussian mixture of experts.
1 code implementation • 29 Jan 2024 • Khai Nguyen, Shujian Zhang, Tam Le, Nhat Ho
From the RPD, we derive the random-path slicing distribution (RPSD) and two variants of sliced Wasserstein, i. e., the Random-Path Projection Sliced Wasserstein (RPSW) and the Importance Weighted Random-Path Projection Sliced Wasserstein (IWRPSW).
1 code implementation • 21 Sep 2023 • Khai Nguyen, Nicola Bariletto, Nhat Ho
Monte Carlo (MC) integration has been employed as the standard approximation method for the Sliced Wasserstein (SW) distance, whose analytical expression involves an intractable expectation.
no code implementations • 27 May 2023 • Tung Le, Khai Nguyen, Shanlin Sun, Kun Han, Nhat Ho, Xiaohui Xie
The metric is defined by sliced Wasserstein distance on meshes represented as probability measures that generalize the set-based approach.
1 code implementation • 12 May 2023 • Huy Nguyen, TrungTin Nguyen, Khai Nguyen, Nhat Ho
Originally introduced as a neural network for ensemble learning, mixture of experts (MoE) has recently become a fundamental building block of highly successful modern deep neural networks for heterogeneous data analysis in several applications of machine learning and statistics.
1 code implementation • 30 Apr 2023 • Khai Nguyen, Nhat Ho
To bridge the literature on variance reduction and the literature on the SW distance, we propose computationally efficient control variates to reduce the variance of the empirical estimation of the SW distance.
1 code implementation • NeurIPS 2023 • Khai Nguyen, Nhat Ho
The second approach is optimizing for the best distribution which belongs to a parametric family of distributions and can maximize the expected distance.
1 code implementation • 12 Jan 2023 • Khai Nguyen, Dang Nguyen, Nhat Ho
Despite being efficient, Max-SW and its amortized version cannot guarantee metricity property due to the sub-optimality of the projected gradient ascent and the amortization gap.
1 code implementation • NeurIPS 2023 • Khai Nguyen, Tongzheng Ren, Nhat Ho
Sliced Wasserstein (SW) distance suffers from redundant projections due to independent uniform random projecting directions.
no code implementations • 19 Oct 2022 • Dung Le, Huy Nguyen, Khai Nguyen, Trang Nguyen, Nhat Ho
Generalized sliced Wasserstein distance is a variant of sliced Wasserstein distance that exploits the power of non-linear projection through a given defining function to better capture the complex structures of the probability distributions.
1 code implementation • 27 Sep 2022 • Khai Nguyen, Tongzheng Ren, Huy Nguyen, Litu Rout, Tan Nguyen, Nhat Ho
We explain the usage of these projections by introducing Hierarchical Radon Transform (HRT) which is constructed by applying Radon Transform variants recursively.
no code implementations • 1 Jun 2022 • Tan Nguyen, Minh Pham, Tam Nguyen, Khai Nguyen, Stanley J. Osher, Nhat Ho
Multi-head attention empowers the recent success of transformers, the state-of-the-art models that have achieved remarkable success in sequence modeling and beyond.
2 code implementations • 4 Apr 2022 • Khai Nguyen, Nhat Ho
Finally, we demonstrate the favorable performance of CSW over the conventional sliced Wasserstein in comparing probability measures over images and in training deep generative modeling on images.
1 code implementation • 25 Mar 2022 • Khai Nguyen, Nhat Ho
Seeking informative projecting directions has been an important task in utilizing sliced Wasserstein distance in applications.
no code implementations • 29 Oct 2021 • Dang Nguyen, Trang Nguyen, Khai Nguyen, Dinh Phung, Hung Bui, Nhat Ho
To address this issue, we propose a novel model fusion framework, named CLAFusion, to fuse neural networks with a different number of layers, which we refer to as heterogeneous neural networks, via cross-layer alignment.
2 code implementations • 22 Aug 2021 • Khai Nguyen, Dang Nguyen, The-Anh Vu-Le, Tung Pham, Nhat Ho
Mini-batch optimal transport (m-OT) has been widely used recently to deal with the memory issue of OT in large-scale applications.
no code implementations • NeurIPS 2021 • Son Nguyen, Duong Nguyen, Khai Nguyen, Khoat Than, Hung Bui, Nhat Ho
Approximate inference in Bayesian deep networks exhibits a dilemma of how to yield high fidelity posterior approximations while maintaining computational efficiency and scalability.
2 code implementations • 11 Feb 2021 • Khai Nguyen, Dang Nguyen, Quoc Nguyen, Tung Pham, Hung Bui, Dinh Phung, Trung Le, Nhat Ho
To address these problems, we propose a novel mini-batch scheme for optimal transport, named Batch of Mini-batches Optimal Transport (BoMb-OT), that finds the optimal coupling between mini-batches and it can be seen as an approximation to a well-defined distance on the space of probability measures.
2 code implementations • ICLR 2021 • Khai Nguyen, Son Nguyen, Nhat Ho, Tung Pham, Hung Bui
To improve the discrepancy and consequently the relational regularization, we propose a new relational discrepancy, named spherical sliced fused Gromov Wasserstein (SSFG), that can find an important area of projections characterized by a von Mises-Fisher distribution.
1 code implementation • ICLR 2021 • Khai Nguyen, Nhat Ho, Tung Pham, Hung Bui
Sliced-Wasserstein distance (SW) and its variant, Max Sliced-Wasserstein distance (Max-SW), have been used widely in the recent years due to their fast computation and scalability even when the probability measures lie in a very high dimensional space.
no code implementations • 26 Jun 2018 • Phuc Nguyen, Khai Nguyen, Ryutaro Ichise, Hideaki Takeda
Semantic labeling for numerical values is a task of assigning semantic labels to unknown numerical attributes.