Search Results for author: Xing Han

Found 19 papers, 7 papers with code

Multi-Pair Text Style Transfer for Unbalanced Data via Task-Adaptive Meta-Learning

no code implementations ACL (MetaNLP) 2021 Xing Han, Jessica Lundin

Text-style transfer aims to convert text given in one domain into another by paraphrasing the sentence or substituting the keywords without altering the content.

Meta-Learning Sentence +2

WATCH: Adaptive Monitoring for AI Deployments via Weighted-Conformal Martingales

1 code implementation7 May 2025 Drew Prinster, Xing Han, Anqi Liu, Suchi Saria

Responsibly deploying artificial intelligence (AI) / machine learning (ML) systems in high-stakes settings arguably requires not only proof of system reliability, but moreover continual, post-deployment monitoring to quickly detect and address any unsafe behavior.

Change Point Detection

Between Linear and Sinusoidal: Rethinking the Time Encoder in Dynamic Graph Learning

1 code implementation10 Apr 2025 Hsing-Huan Chung, Shravan Chaudhari, Xing Han, Yoav Wald, Suchi Saria, Joydeep Ghosh

Through extensive experiments on six dynamic graph datasets, we demonstrate that the linear time encoder improves the performance of TGAT and DyGFormer in most cases.

Graph Learning

Quadratic Gating Functions in Mixture of Experts: A Statistical Insight

no code implementations15 Oct 2024 Pedram Akbarian, Huy Nguyen, Xing Han, Nhat Ho

Mixture of Experts (MoE) models are highly effective in scaling model capacity while preserving computational efficiency, with the gating network, or router, playing a central role by directing inputs to the appropriate experts.

Computational Efficiency Mixture-of-Experts +1

On Expert Estimation in Hierarchical Mixture of Experts: Beyond Softmax Gating Functions

no code implementations3 Oct 2024 Huy Nguyen, Xing Han, Carl William Harris, Suchi Saria, Nhat Ho

With the growing prominence of the Mixture of Experts (MoE) architecture in developing large-scale foundation models, we investigate the Hierarchical Mixture of Experts (HMoE), a specialized variant of MoE that excels in handling complex inputs and improving performance on targeted tasks.

image-classification Image Classification +1

Optimizing RAG Techniques for Automotive Industry PDF Chatbots: A Case Study with Locally Deployed Ollama Models

no code implementations12 Aug 2024 Fei Liu, Zejun Kang, Xing Han

To evaluate our approach, we constructed a proprietary dataset comprising typical automotive industry documents, including technical reports and corporate regulations.

RAG Retrieval +1

Achieving Fairness Across Local and Global Models in Federated Learning

no code implementations24 Jun 2024 Disha Makhija, Xing Han, Joydeep Ghosh, Yejin Kim

Achieving fairness across diverse clients in Federated Learning (FL) remains a significant challenge due to the heterogeneity of the data and the inaccessibility of sensitive attributes from clients' private datasets.

Fairness Federated Learning

Novel Node Category Detection Under Subpopulation Shift

2 code implementations1 Apr 2024 Hsing-Huan Chung, Shravan Chaudhari, Yoav Wald, Xing Han, Joydeep Ghosh

We introduce a new approach, Recall-Constrained Optimization with Selective Link Prediction (RECO-SLIP), to detect nodes belonging to novel categories in attributed graphs under subpopulation shifts.

Link Prediction

LSTTN: A Long-Short Term Transformer-based Spatio-temporal Neural Network for Traffic Flow Forecasting

1 code implementation25 Mar 2024 Qinyao Luo, Silu He, Xing Han, YuHan Wang, Haifeng Li

Accurate traffic forecasting is a fundamental problem in intelligent transportation systems and learning long-range traffic representations with key information through spatiotemporal graph neural networks (STGNNs) is a basic assumption of current traffic flow prediction models.

FuseMoE: Mixture-of-Experts Transformers for Fleximodal Fusion

no code implementations5 Feb 2024 Xing Han, Huy Nguyen, Carl Harris, Nhat Ho, Suchi Saria

As machine learning models in critical fields increasingly grapple with multimodal data, they face the dual challenges of handling a wide array of modalities, often incomplete due to missing elements, and the temporal irregularity and sparsity of collected samples.

Missing Elements Mixture-of-Experts

Split Localized Conformal Prediction

1 code implementation27 Jun 2022 Xing Han, Ziyang Tang, Joydeep Ghosh, Qiang Liu

The modified score inherits the spirit of split conformal methods, which is simple and efficient and can scale to high dimensional settings.

Conformal Prediction Density Estimation +3

Efficient Forecasting of Large Scale Hierarchical Time Series via Multilevel Clustering

no code implementations27 May 2022 Xing Han, Tongzheng Ren, Jing Hu, Joydeep Ghosh, Nhat Ho

To attain this goal, each time series is first assigned the forecast for its cluster representative, which can be considered as a "shrinkage prior" for the set of time series it represents.

Clustering Time Series +1

Architecture Agnostic Federated Learning for Neural Networks

no code implementations15 Feb 2022 Disha Makhija, Xing Han, Nhat Ho, Joydeep Ghosh

With growing concerns regarding data privacy and rapid increase in data volume, Federated Learning(FL) has become an important learning paradigm.

Federated Learning

Dynamic Combination of Heterogeneous Models for Hierarchical Time Series

no code implementations22 Dec 2021 Xing Han, Jing Hu, Joydeep Ghosh

We conduct a comprehensive evaluation of both point and quantile forecasts for hierarchical time series (HTS), including public data and user records from a large financial software company.

Time Series Time Series Analysis

MECATS: Mixture-of-Experts for Probabilistic Forecasts of Aggregated Time Series

no code implementations29 Sep 2021 Xing Han, Jing Hu, Joydeep Ghosh

We introduce a mixture of heterogeneous experts framework called MECATS, which simultaneously forecasts the values of a set of time series that are related through an aggregation hierarchy.

Mixture-of-Experts Time Series +1

Multi-Pair Text Style Transfer on Unbalanced Data

no code implementations20 Jun 2021 Xing Han, Jessica Lundin

Text-style transfer aims to convert text given in one domain into another by paraphrasing the sentence or substituting the keywords without altering the content.

Meta-Learning Sentence +2

Simultaneously Reconciled Quantile Forecasting of Hierarchically Related Time Series

1 code implementation25 Feb 2021 Xing Han, Sambarta Dasgupta, Joydeep Ghosh

In such applications, it is important that the forecasts, in addition to being reasonably accurate, are also consistent w. r. t one another.

quantile regression Time Series +1

Certified Monotonic Neural Networks

1 code implementation NeurIPS 2020 Xingchao Liu, Xing Han, Na Zhang, Qiang Liu

In this work, we propose to certify the monotonicity of the general piece-wise linear neural networks by solving a mixed integer linear programming problem. This provides a new general approach for learning monotonic neural networks with arbitrary model structures.

Fairness

Model-Agnostic Explanations using Minimal Forcing Subsets

no code implementations1 Nov 2020 Xing Han, Joydeep Ghosh

How can we find a subset of training samples that are most responsible for a specific prediction made by a complex black-box machine learning model?

BIG-bench Machine Learning Counterfactual Explanation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.