Search Results for author: Zhiyuan Fan

Found 5 papers, 1 papers with code

Settling Constant Regrets in Linear Markov Decision Processes

no code implementations16 Apr 2024 Weitong Zhang, Zhiyuan Fan, Jiafan He, Quanquan Gu

To the best of our knowledge, Cert-LSVI-UCB is the first algorithm to achieve a constant, instance-dependent, high-probability regret bound in RL with linear function approximation for infinite runs without relying on prior distribution assumptions.

Reinforcement Learning (RL)

Efficient Data Learning for Open Information Extraction with Pre-trained Language Models

no code implementations23 Oct 2023 Zhiyuan Fan, Shizhu He

Open Information Extraction (OpenIE) is a fundamental yet challenging task in Natural Language Processing, which involves extracting all triples (subject, predicate, object) from a given sentence.

Open Information Extraction Sentence

On the Interplay Between Misspecification and Sub-optimality Gap in Linear Contextual Bandits

no code implementations16 Mar 2023 Weitong Zhang, Jiafan He, Zhiyuan Fan, Quanquan Gu

We show that, when the misspecification level $\zeta$ is dominated by $\tilde O (\Delta / \sqrt{d})$ with $\Delta$ being the minimal sub-optimality gap and $d$ being the dimension of the contextual vectors, our algorithm enjoys the same gap-dependent regret bound $\tilde O (d^2/\Delta)$ as in the well-specified setting up to logarithmic factors.

Multi-Armed Bandits

OpenFE: Automated Feature Generation with Expert-level Performance

2 code implementations22 Nov 2022 Tianping Zhang, Zheyu Zhang, Zhiyuan Fan, Haoyan Luo, Fengyuan Liu, Qian Liu, Wei Cao, Jian Li

In the two competitions, features generated by OpenFE with a simple baseline model can beat 99. 3% and 99. 6% data science teams respectively.

Feature Importance

Efficient Algorithms for Sparse Moment Problems without Separation

no code implementations26 Jul 2022 Zhiyuan Fan, Jian Li

Our algorithm for the one-dimensional problem (also called the sparse Hausdorff moment problem) is a robust version of the classic Prony's method, and our contribution mainly lies in the analysis.

Topic Models

Cannot find the paper you are looking for? You can Submit a new open access paper.