no code implementations • 22 May 2025 • Korel Gundem, Juncheng Dong, Dennis Zhang, Vahid Tarokh, Zhengling Qi
While calibration techniques are proposed to mitigate these biases, we show that, in the logit space, many of these methods are equivalent to merely shifting the LLM's decision boundary without having the ability to alter its orientation.
no code implementations • 17 Feb 2025 • Zihao Wu, Juncheng Dong, Ahmed Aloui, Vahid Tarokh
Optimization techniques have become increasingly critical due to the ever-growing model complexity and data scale.
no code implementations • 17 Feb 2025 • Zihao Wu, Juncheng Dong, Haoming Yang, Vahid Tarokh
Time series forecasting has recently achieved significant progress with multi-scale models to address the heterogeneity between long and short range patterns.
no code implementations • 31 Dec 2024 • Ahmed Aloui, Ali Hasan, Juncheng Dong, Zihao Wu, Vahid Tarokh
In this paper, we introduce a new approach for integrating score-based models with the Metropolis-Hastings algorithm.
no code implementations • 3 Dec 2024 • Juncheng Dong, Zihao Wu, Hamid Jafarkhani, Ali Pezeshki, Vahid Tarokh
Many challenges in science and engineering, such as drug discovery and communication network design, involve optimizing complex and expensive black-box functions across vast search spaces.
no code implementations • 7 Nov 2023 • Ahmed Aloui, Juncheng Dong, Cat P. Le, Vahid Tarokh
To address this issue, we propose a model-agnostic data augmentation method for CATE estimation.
no code implementations • 12 Jun 2023 • Juncheng Dong, Hao-Lun Hsu, Qitong Gao, Vahid Tarokh, Miroslav Pajic
In this work, we extend the two-player game by introducing an adversarial herd, which involves a group of adversaries, in order to address ($\textit{i}$) the difficulty of the inner optimization problem, and ($\textit{ii}$) the potential over pessimism caused by the selection of a candidate adversary set that may include unlikely scenarios.
no code implementations • 19 May 2023 • Cat P. Le, Juncheng Dong, Ahmed Aloui, Vahid Tarokh
To this end, we introduce a new continual learning approach for conditional generative adversarial networks by leveraging a mode-affinity score specifically designed for generative modeling.
no code implementations • 8 Feb 2023 • Juncheng Dong, Weibin Mo, Zhengling Qi, Cong Shi, Ethan X. Fang, Vahid Tarokh
The objective is to use the offline dataset to find an optimal assortment.
no code implementations • 3 Feb 2023 • Yiling Liu, Juncheng Dong, Ziyang Jiang, Ahmed Aloui, Keyu Li, Hunter Klein, Vahid Tarokh, David Carlson
To address this limitation, we propose a novel generalization bound that reweights source classification error by aligning source and target sub-domains.
no code implementations • 1 Oct 2022 • Ahmed Aloui, Juncheng Dong, Cat P. Le, Vahid Tarokh
To this end, we theoretically assess the feasibility of transferring ITE knowledge and present a practical framework for efficient transfer.
no code implementations • 22 Jan 2022 • Juncheng Dong, Suya Wu, Mohammadreza Sultani, Vahid Tarokh
In particular, by modeling the adversaries as learning agents, we show that the proposed MAAS is able to successfully choose the transmitted channel(s) and their respective allocated power(s) without any prior knowledge of the sender strategy.
no code implementations • ICLR 2022 • Juncheng Dong, Simiao Ren, Yang Deng, Omar Khatib, Jordan Malof, Mohammadreza Soltani, Willie Padilla, Vahid Tarokh
To this end, we propose a physics-infused deep neural network based on the Blaschke products for phase retrieval.
1 code implementation • ICLR 2022 • Cat P. Le, Juncheng Dong, Mohammadreza Soltani, Vahid Tarokh
We propose an asymmetric affinity score for representing the complexity of utilizing the knowledge of one task for learning another one.
1 code implementation • 23 Mar 2021 • Cat P. Le, Mohammadreza Soltani, Juncheng Dong, Vahid Tarokh
Next, we construct an online neural architecture search framework using the Fisher task distance, in which we have access to the past learned tasks.