no code implementations • 16 Mar 2024 • Mariia Khan, Yue Qiu, Yuren Cong, Jumana Abu-Khalaf, David Suter, Bodo Rosenhahn
The foundational Segment Anything Model (SAM) is designed for promptable multi-class multi-instance segmentation but tends to output part or sub-part masks in the "everything" mode for various real-world applications.
no code implementations • 1 Feb 2024 • Jianguo Huang, Yue Qiu
Neural operators (NO) are discretization invariant deep learning methods with functional output and can approximate any continuous operator.
no code implementations • 22 Jan 2024 • Yuhuang Meng, Yue Qiu
Sparse identification of differential equations aims to compute the analytic expressions from the observed data explicitly.
2 code implementations • 10 Oct 2023 • Jianguo Huang, Huajun Xi, Linjun Zhang, Huaxiu Yao, Yue Qiu, Hongxin Wei
In this paper, we empirically and theoretically show that disregarding the probabilities' value will mitigate the undesirable effect of miscalibrated probability values.
no code implementations • 30 Jun 2023 • Yuhuang Meng, Jianguo Huang, Yue Qiu
In Koopman operator theory, a finite-dimensional nonlinear system is transformed into an infinite but linear system using a set of observable functions.
no code implementations • 18 May 2023 • Jiankui Zhou, Yue Qiu
Stein Variational Gradient Descent (SVGD) is a popular particle-based method for Bayesian inference.
1 code implementation • 16 Apr 2023 • Leyuan Sun, Guanqun Ding, Yue Qiu, Yusuke Yoshiyasu, Fumio Kanehiro
A synthetic multi-modal dataset is made public to validate the generalization ability of the proposed fusion strategy, which also works for other combinations of different modalities.
no code implementations • CVPR 2023 • Yue Qiu, Yanjun Sun, Fumiya Matsuzawa, Kenji Iwata, Hirokatsu Kataoka
This paper proposes a new visual reasoning formulation that aims at discovering changes between image pairs and their temporal orders.
2 code implementations • ICCV 2021 • Yue Qiu, Shintaro Yamamoto, Kodai Nakashima, Ryota Suzuki, Kenji Iwata, Hirokatsu Kataoka, Yutaka Satoh
Change captioning tasks aim to detect changes in image pairs observed before and after a scene change and generate a natural language description of the changes.