Search Results for author: Kaixin Wang

Found 20 papers, 9 papers with code

How Far is Video Generation from World Model: A Physical Law Perspective

no code implementations4 Nov 2024 Bingyi Kang, Yang Yue, Rui Lu, Zhijie Lin, Yang Zhao, Kaixin Wang, Gao Huang, Jiashi Feng

Our scaling experiments show perfect generalization within the distribution, measurable scaling behavior for combinatorial generalization, but failure in out-of-distribution scenarios.

Video Generation

Large Language Model-Based Agents for Software Engineering: A Survey

1 code implementation4 Sep 2024 Junwei Liu, Kaixin Wang, Yixuan Chen, Xin Peng, Zhenpeng Chen, Lingming Zhang, Yiling Lou

The recent advance in Large Language Models (LLMs) has shaped a new paradigm of AI agents, i. e., LLM-based agents.

AI Agent Language Modelling +1

Vul-RAG: Enhancing LLM-based Vulnerability Detection via Knowledge-level RAG

no code implementations17 Jun 2024 Xueying Du, Geng Zheng, Kaixin Wang, Jiayi Feng, Wentai Deng, Mingwei Liu, Bihuan Chen, Xin Peng, Tao Ma, Yiling Lou

In addition, our user study shows that the vulnerability knowledge generated by Vul-RAG can serve as high-quality explanations which can improve the manual detection accuracy from 0. 60 to 0. 77.

RAG Vulnerability Detection

Improving Token-Based World Models with Parallel Observation Prediction

1 code implementation8 Feb 2024 Lior Cohen, Kaixin Wang, Bingyi Kang, Shie Mannor

We incorporate POP in a novel TBWM agent named REM (Retentive Environment Model), showcasing a 15. 4x faster imagination compared to prior TBWMs.

ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation

1 code implementation3 Aug 2023 Xueying Du, Mingwei Liu, Kaixin Wang, Hanlin Wang, Junwei Liu, Yixuan Chen, Jiayi Feng, Chaofeng Sha, Xin Peng, Yiling Lou

Third, we find that generating the entire class all at once (i. e. holistic generation strategy) is the best generation strategy only for GPT-4 and GPT-3. 5, while method-by-method generation (i. e. incremental and compositional) is better strategies for the other models with limited ability of understanding long instructions and utilizing the middle information.

Class-level Code Generation HumanEval

Bring Your Own (Non-Robust) Algorithm to Solve Robust MDPs by Estimating The Worst Kernel

no code implementations9 Jun 2023 Kaixin Wang, Uri Gadot, Navdeep Kumar, Kfir Levy, Shie Mannor

Robust Markov Decision Processes (RMDPs) provide a framework for sequential decision-making that is robust to perturbations on the transition kernel.

Decision Making reinforcement-learning +2

An Efficient Solution to s-Rectangular Robust Markov Decision Processes

no code implementations31 Jan 2023 Navdeep Kumar, Kfir Levy, Kaixin Wang, Shie Mannor

We present an efficient robust value iteration for \texttt{s}-rectangular robust Markov Decision Processes (MDPs) with a time complexity comparable to standard (non-robust) MDPs which is significantly faster than any existing method.

LEMMA

Reinforcement Learning Enhanced Weighted Sampling for Accurate Subgraph Counting on Fully Dynamic Graph Streams

1 code implementation13 Nov 2022 Kaixin Wang, Cheng Long, Da Yan, Jie Zhang, H. V. Jagadish

Specifically, we propose a weighted sampling algorithm called WSD for estimating the subgraph count in a fully dynamic graph stream, which samples the edges based on their weights that indicate their importance and reflect their properties.

Subgraph Counting

Reachability-Aware Laplacian Representation in Reinforcement Learning

no code implementations24 Oct 2022 Kaixin Wang, Kuangqi Zhou, Jiashi Feng, Bryan Hooi, Xinchao Wang

In Reinforcement Learning (RL), Laplacian Representation (LapRep) is a task-agnostic state representation that encodes the geometry of the environment.

reinforcement-learning Reinforcement Learning +1

Policy Gradient for Reinforcement Learning with General Utilities

no code implementations3 Oct 2022 Navdeep Kumar, Kaixin Wang, Kfir Levy, Shie Mannor

The policy gradient theorem proves to be a cornerstone in Linear RL due to its elegance and ease of implementability.

reinforcement-learning Reinforcement Learning +1

Relational Reasoning via Set Transformers: Provable Efficiency and Applications to MARL

no code implementations20 Sep 2022 Fengzhuo Zhang, Boyi Liu, Kaixin Wang, Vincent Y. F. Tan, Zhuoran Yang, Zhaoran Wang

The cooperative Multi-A gent R einforcement Learning (MARL) with permutation invariant agents framework has achieved tremendous empirical successes in real-world applications.

Relational Reasoning

Efficient Policy Iteration for Robust Markov Decision Processes via Regularization

1 code implementation28 May 2022 Navdeep Kumar, Kfir Levy, Kaixin Wang, Shie Mannor

But we don't have a clear understanding to exploit this equivalence, to do policy improvement steps to get the optimal value function or policy.

Tyger: Task-Type-Generic Active Learning for Molecular Property Prediction

no code implementations23 May 2022 Kuangqi Zhou, Kaixin Wang, Jiashi Feng, Jian Tang, Tingyang Xu, Xinchao Wang

However, existing best deep AL methods are mostly developed for a single type of learning task (e. g., single-label classification), and hence may not perform well in molecular property prediction that involves various task types.

Active Learning Drug Discovery +3

The Geometry of Robust Value Functions

no code implementations30 Jan 2022 Kaixin Wang, Navdeep Kumar, Kuangqi Zhou, Bryan Hooi, Jiashi Feng, Shie Mannor

The key of this perspective is to decompose the value space, in a state-wise manner, into unions of hypersurfaces.

Improving Generalization in Reinforcement Learning with Mixture Regularization

2 code implementations NeurIPS 2020 Kaixin Wang, Bingyi Kang, Jie Shao, Jiashi Feng

Deep reinforcement learning (RL) agents trained in a limited set of environments tend to suffer overfitting and fail to generalize to unseen testing environments.

Data Augmentation Deep Reinforcement Learning +3

Understanding and Resolving Performance Degradation in Graph Convolutional Networks

2 code implementations12 Jun 2020 Kuangqi Zhou, Yanfei Dong, Kaixin Wang, Wee Sun Lee, Bryan Hooi, Huan Xu, Jiashi Feng

In this work, we study performance degradation of GCNs by experimentally examining how stacking only TRANs or PROPs works.

PANet: Few-Shot Image Semantic Segmentation with Prototype Alignment

5 code implementations ICCV 2019 Kaixin Wang, Jun Hao Liew, Yingtian Zou, Daquan Zhou, Jiashi Feng

In this paper, we tackle the challenging few-shot segmentation problem from a metric learning perspective and present PANet, a novel prototype alignment network to better utilize the information of the support set.

Few-Shot Semantic Segmentation Metric Learning +2

Neural Epitome Search for Architecture-Agnostic Network Compression

no code implementations ICLR 2020 Daquan Zhou, Xiaojie Jin, Qibin Hou, Kaixin Wang, Jianchao Yang, Jiashi Feng

The recent WSNet [1] is a new model compression method through sampling filterweights from a compact set and has demonstrated to be effective for 1D convolutionneural networks (CNNs).

Model Compression Neural Architecture Search

Cannot find the paper you are looking for? You can Submit a new open access paper.