Search Results for author: Zhao Yang

Found 38 papers, 15 papers with code

Regulatory DNA sequence Design with Reinforcement Learning

1 code implementation11 Mar 2025 Zhao Yang, Bing Su, Chuan Cao, Ji-Rong Wen

The fitness of CREs, measured by their ability to modulate gene expression, highly depends on the nucleotide sequences, especially specific motifs known as transcription factor binding sites (TFBSs).

reinforcement-learning Reinforcement Learning +1

DualDiff+: Dual-Branch Diffusion for High-Fidelity Video Generation with Reward Guidance

1 code implementation5 Mar 2025 Zhao Yang, Zezhong Qian, Xiaofan Li, Weixiang Xu, Gongpeng Zhao, Ruohong Yu, Lingsi Zhu, Longjun Liu

In this work, we present DualDiff, a dual-branch conditional diffusion model designed to enhance driving scene generation across multiple views and video sequences.

3D Object Detection BEV Segmentation +3

HybriDNA: A Hybrid Transformer-Mamba2 Long-Range DNA Language Model

no code implementations15 Feb 2025 Mingqian Ma, Guoqing Liu, Chuan Cao, Pan Deng, Tri Dao, Albert Gu, Peiran Jin, Zhao Yang, Yingce Xia, Renqian Luo, Pipi Hu, Zun Wang, Yuan-Jyue Chen, Haiguang Liu, Tao Qin

To address these challenges, we propose HybriDNA, a decoder-only DNA language model that incorporates a hybrid Transformer-Mamba2 architecture, seamlessly integrating the strengths of attention mechanisms with selective state-space models.

Language Modeling Language Modelling +1

Fact-Preserved Personalized News Headline Generation

1 code implementation IEEE International Conference on Data Mining (ICDM) 2023 Zhao Yang, Junhong Lian, Xiang Ao

In this paper, we propose a framework Fact-Preserved Personalized News Headline Generation (short for FPG), to prompt a tradeoff between personalization and consistency.

Contrastive Learning Headline Generation

Interpretable Enzyme Function Prediction via Residue-Level Detection

1 code implementation10 Jan 2025 Zhao Yang, Bing Su, Jiahao Chen, Ji-Rong Wen

It uses a set of learnable functional queries to adaptatively extract different local representations from the sequence of residue-level features for predicting different EC numbers.

Multi-Label Classification MUlTI-LABEL-ClASSIFICATION +1

World Models Increase Autonomy in Reinforcement Learning

no code implementations19 Aug 2024 Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat, Edward S. Hu

However, the training process of RL is far from automatic, requiring extensive human effort to reset the agent and environments.

reinforcement-learning Reinforcement Learning +1

Improving Zero-shot LLM Re-Ranker with Risk Minimization

no code implementations19 Jun 2024 Xiaowei Yuan, Zhao Yang, Yequan Wang, Jun Zhao, Kang Liu

In the Retrieval-Augmented Generation (RAG) system, advanced Large Language Models (LLMs) have emerged as effective Query Likelihood Models (QLMs) in an unsupervised way, which re-rank documents based on the probability of generating the query given the content of a document.

RAG Re-Ranking +1

Dual-Tap Optical-Digital Feedforward Equalization Enabling High-Speed Optical Transmission in IM/DD Systems

no code implementations1 Feb 2024 Yu Guo, Yangbo Wu, Zhao Yang, Lei Xue, Ning Liang, Yang Ren, Zhengrui Tu, Jia Feng, Qunbi Zhuge

Intensity-modulation and direct-detection (IM/DD) transmission is widely adopted for high-speed optical transmission scenarios due to its cost-effectiveness and simplicity.

CaKDP: Category-aware Knowledge Distillation and Pruning Framework for Lightweight 3D Object Detection

1 code implementation CVPR 2024 Haonan Zhang, Longjun Liu, Yuqi Huang, Zhao Yang, Xinyu Lei, Bihan Wen

To address these issues we propose a simple yet effective Category-aware Knowledge Distillation and Pruning (CaKDP) framework for compressing 3D detectors.

3D Object Detection Knowledge Distillation +1

AppAgent: Multimodal Agents as Smartphone Users

1 code implementation21 Dec 2023 Chi Zhang, Zhao Yang, Jiaxuan Liu, Yucheng Han, Xin Chen, Zebiao Huang, Bin Fu, Gang Yu

Recent advancements in large language models (LLMs) have led to the creation of intelligent agents capable of performing complex tasks.

Navigate

Synthesizing Long-Term Human Motions with Diffusion Models via Coherent Sampling

1 code implementation3 Aug 2023 Zhao Yang, Bing Su, Ji-Rong Wen

Firstly, they cannot directly generate coherent motions and require additional operations such as interpolation to process the generated actions.

Motion Generation Sentence

Two-Memory Reinforcement Learning

no code implementations20 Apr 2023 Zhao Yang, Thomas. M. Moerland, Mike Preuss, Aske Plaat

While deep reinforcement learning has shown important empirical success, it tends to learn relatively slow due to slow propagation of rewards information and slow update of parametric neural networks.

Deep Reinforcement Learning reinforcement-learning +2

First Go, then Post-Explore: the Benefits of Post-Exploration in Intrinsic Motivation

no code implementations6 Dec 2022 Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat

In this paper, we present a clear ablation study of post-exploration in a general intrinsically motivated goal exploration process (IMGEP) framework, that the Go-Explore paper did not show.

continuous-control Continuous Control +2

Continuous Episodic Control

no code implementations28 Nov 2022 Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat

Therefore, this paper introduces Continuous Episodic Control (CEC), a novel non-parametric episodic memory algorithm for sequential decision making in problems with a continuous action space.

continuous-control Continuous Control +5

A Molecular Multimodal Foundation Model Associating Molecule Graphs with Natural Language

4 code implementations12 Sep 2022 Bing Su, Dazhao Du, Zhao Yang, Yujie Zhou, Jiangmeng Li, Anyi Rao, Hao Sun, Zhiwu Lu, Ji-Rong Wen

Although artificial intelligence (AI) has made significant progress in understanding molecules in a wide range of fields, existing models generally acquire the single cognitive ability from the single molecular modality.

Contrastive Learning Cross-Modal Retrieval +4

When to Go, and When to Explore: The Benefit of Post-Exploration in Intrinsic Motivation

no code implementations29 Mar 2022 Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat

Go-Explore achieved breakthrough performance on challenging reinforcement learning (RL) tasks with sparse rewards.

Reinforcement Learning (RL)

On the Effectiveness of Pinyin-Character Dual-Decoding for End-to-End Mandarin Chinese ASR

no code implementations26 Jan 2022 Zhao Yang, Dianwen Ng, Xiao Fu, Liping Han, Wei Xi, Rui Wang, Rui Jiang, Jizhong Zhao

Based on the above intuition, we first investigate types of end-to-end encoder-decoder based models in the single-input dual-output (SIDO) multi-task framework, after which a novel asynchronous decoding with fuzzy Pinyin sampling method is proposed according to the one-to-one correspondence characteristics between Pinyin and Character.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

Logic Traps in Evaluating Attribution Scores

no code implementations ACL 2022 Yiming Ju, Yuanzhe Zhang, Zhao Yang, Zhongtao Jiang, Kang Liu, Jun Zhao

Meanwhile, since the reasoning process of deep models is inaccessible, researchers design various evaluation methods to demonstrate their arguments.

Potential-based Reward Shaping in Sokoban

no code implementations10 Sep 2021 Zhao Yang, Mike Preuss, Aske Plaat

While previous work has investigated the use of expert knowledge to generate potential functions, in this work, we study whether we can use a search algorithm(A*) to automatically generate a potential function for reward shaping in Sokoban, a well-known planning task.

Sokoban

Follow the Prophet: Accurate Online Conversion Rate Prediction in the Face of Delayed Feedback

1 code implementation13 Aug 2021 Haoming Li, Feiyang Pan, Xiang Ao, Zhao Yang, Min Lu, Junwei Pan, Dapeng Liu, Lei Xiao, Qing He

The delayed feedback problem is one of the imperative challenges in online advertising, which is caused by the highly diversified feedback delay of a conversion varying from a few minutes to several days.

Alignment Rationale for Natural Language Inference

no code implementations ACL 2021 Zhongtao Jiang, Yuanzhe Zhang, Zhao Yang, Jun Zhao, Kang Liu

Deep learning models have achieved great success on the task of Natural Language Inference (NLI), though only a few attempts try to explain their behaviors.

feature selection Natural Language Inference

Transfer Learning and Curriculum Learning in Sokoban

no code implementations25 May 2021 Zhao Yang, Mike Preuss, Aske Plaat

In reinforcement learning, learning actions for a behavior policy that can be applied to new environments is still a challenge, especially for tasks that involve much planning.

reinforcement-learning Reinforcement Learning +3

Explore User Neighborhood for Real-time E-commerce Recommendation

no code implementations28 Feb 2021 Xu Xie, Fei Sun, Xiaoyong Yang, Zhao Yang, Jinyang Gao, Wenwu Ou, Bin Cui

On the one hand, it utilizes UI relations and user neighborhood to capture both global and local information.

Collaborative Filtering Recommendation Systems

Helios: Heterogeneity-Aware Federated Learning with Dynamically Balanced Collaboration

no code implementations3 Dec 2019 Zirui Xu, Zhao Yang, JinJun Xiong, Jianlei Yang, Xiang Chen

In this paper, we propose Helios, a heterogeneity-aware FL framework to tackle the straggler issue.

Distributed, Parallel, and Cluster Computing

Hetero-Center Loss for Cross-Modality Person Re-Identification

no code implementations22 Oct 2019 Yuanxin Zhu, Zhao Yang, Li Wang, Sai Zhao, Xiao Hu, Dapeng Tao

With the joint supervision of Cross-Entropy (CE) loss and HC loss, the network is trained to achieve two vital objectives, inter-class discrepancy and intra-class cross-modality similarity as much as possible.

Cross-Modality Person Re-identification Person Re-Identification

Res-embedding for Deep Learning Based Click-Through Rate Prediction Modeling

no code implementations25 Jun 2019 Guorui Zhou, Kailun Wu, Weijie Bian, Zhao Yang, Xiaoqiang Zhu, Kun Gai

In this paper, we model user behavior using an interest delay model, study carefully the embedding mechanism, and obtain two important results: (i) We theoretically prove that small aggregation radius of embedding vectors of items which belongs to a same user interest domain will result in good generalization performance of deep CTR model.

Click-Through Rate Prediction

Learn to Interpret Atari Agents

1 code implementation29 Dec 2018 Zhao Yang, Song Bai, Li Zhang, Philip H. S. Torr

Deep reinforcement learning (DeepRL) agents surpass human-level performance in many tasks.

Decision Making Deep Reinforcement Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.