no code implementations • 4 Feb 2025 • Chengrui Li, Yunmiao Wang, Yule Wang, Weihan Li, Dieter Jaeger, Anqi Wu
A fully disentangled variational auto-encoder (VAE) aims to identify disentangled latent components from observations.
1 code implementation • 12 Oct 2024 • Yule Wang, Chengrui Li, Weihan Li, Anqi Wu
To tackle this limitation, our approach, named ``BeNeDiff'', first identifies a fine-grained and disentangled neural subspace using a behavior-informed latent variable model.
no code implementations • 29 Jun 2024 • Weihan Li, Yule Wang, Chengrui Li, Anqi Wu
Understanding and constructing brain communications that capture dynamic communications across multiple regions is fundamental to modern system neuroscience, yet current methods struggle to find time-varying region-level communications or scale to large neural datasets with long recording durations.
1 code implementation • 5 Feb 2024 • Weihan Li, Chengrui Li, Yule Wang, Anqi Wu
Consequently, the model achieves a linear inference cost over time points and provides an interpretable low-dimensional representation, revealing communication directions across brain regions and separating oscillatory communications into different frequency bands.
no code implementations • 2 Feb 2024 • Chengrui Li, Weihan Li, Yule Wang, Anqi Wu
For (1), we propose a new differentiable POGLM, which enables the pathwise gradient estimator, better than the score function gradient estimator used in existing works.
no code implementations • 4 Nov 2023 • Chengrui Li, Yule Wang, Weihan Li, Anqi Wu
Maximizing the log-likelihood is a crucial aspect of learning latent variable models, and variational inference (VI) stands as the commonly adopted method.
1 code implementation • 9 Jun 2023 • Yule Wang, Zijing Wu, Chengrui Li, Anqi Wu
Specifically, the latent dynamics structures of the source domain are first extracted by a diffusion model.
no code implementations • 11 Aug 2022 • Yuxiang Shi, Yue Ding, Bo Chen, YuYang Huang, Yule Wang, Ruiming Tang, Dong Wang
In this paper, we propose a Task aligned Meta-learning based Augmented Graph (TMAG) to address cold-start recommendation.
no code implementations • 28 Sep 2021 • Yunzhe Li, Yue Ding, Bo Chen, Xin Xin, Yule Wang, Yuxiang Shi, Ruiming Tang, Dong Wang
In this paper, we propose a novel time-aware sequential recommendation framework called Social Temporal Excitation Networks (STEN), which introduces temporal point processes to model the fine-grained impact of friends' behaviors on the user s dynamic interests in an event-level direct paradigm.
no code implementations • 27 Sep 2021 • Yule Wang, Xin Xin, Yue Ding, Yunzhe Li, Dong Wang
In detail, we define our item cluster-wise optimization target as the recommender model should balance all item clusters that differ in popularity, thus we set the model learning on each item cluster as a unique optimization objective.
no code implementations • 26 Sep 2021 • Yule Wang, Qiang Luo, Yue Ding, Yunzhe Li, Dong Wang, Hongbo Deng
In this paper, we propose a novel model named DemiNet (short for DEpendency-Aware Multi-Interest Network) to address the above two issues.
no code implementations • 19 Jun 2019 • Rongfang Wang, Jia-Wei Chen, Yule Wang, Licheng Jiao, Mi Wang
In this letter, we proposed a spatial metric learning method to obtain a difference image more robust to the speckle by learning a metric from a set of constraint pairs.