no code implementations • 25 Jun 2025 • Kethmi Hirushini Hettige, Jiahao Ji, Cheng Long, Shili Xiang, Gao Cong, Jingyuan Wang
In this work, we introduce STReason, a novel framework that integrates the reasoning strengths of large language models (LLMs) with the analytical capabilities of spatio-temporal models for multi-task inference and execution.
1 code implementation • 20 May 2025 • Qianxiong Xu, Lanyun Zhu, Xuanyi Liu, Guosheng Lin, Cheng Long, Ziyue Li, Rui Zhao
A simple idea is to encode support foreground (FG) features as memory, with which query FG features are matched and fused.
no code implementations • 16 May 2025 • Qiuyu Zhu, Liang Zhang, Qianxiong Xu, Cheng Long, Jie Zhang
Graph-based retrieval-augmented generation (RAG) enables large language models (LLMs) to incorporate structured knowledge via graph retrieval as contextual input, enhancing more accurate and context-aware reasoning.
2 code implementations • 5 May 2025 • Chenxi Liu, Shaowen Zhou, Qianxiong Xu, Hao Miao, Cheng Long, Ziyue Li, Rui Zhao
However, a fundamental cross-modality gap between time series and LLMs exists, as LLMs are pre-trained on textual corpora and are not inherently optimized for time series.
3 code implementations • 4 May 2025 • Chenxi Liu, Hao Miao, Qianxiong Xu, Shaowen Zhou, Cheng Long, Yan Zhao, Ziyue Li, Rui Zhao
To address this problem, we introduce TimeKD, an efficient MTSF framework that leverages the calibrated language models and privileged knowledge distillation.
Knowledge Distillation
Multivariate Time Series Forecasting
+1
no code implementations • 6 Mar 2025 • Ziqiao Wang, Cheng Long, Yongyi Mao
Federated learning (FL) is a widely adopted privacy-preserving distributed learning framework, yet its generalization performance remains less explored compared to centralized learning.
no code implementations • 22 Jan 2025 • Qiuyu Zhu, Liang Zhang, Qianxiong Xu, Cheng Long
However, this separation fails to capture the critical interactions between these two types of information within HTRNs.
no code implementations • 3 Dec 2024 • Jie Zou, Aixin Sun, Cheng Long, Evangelos Kanoulas
These items and item-related entities are often mentioned along the development of a dialog, leading to potential sequential dependencies among them.
1 code implementation • 20 Nov 2024 • Sixiao Zhang, Cheng Long, Wei Yuan, Hongxu Chen, Hongzhi Yin
In this work, we explore the problem of data watermarking for sequential recommender systems, where a watermark is embedded into the target dataset and can be detected in models trained on that dataset.
2 code implementations • 19 Nov 2024 • Yutong Gou, Jianyang Gao, Yuexuan Xu, Cheng Long
However, they face performance bottlenecks due to the random memory accesses caused by the searching process on the graph indices and the costs of computing exact distances to guide the searching process.
no code implementations • 3 Nov 2024 • Mingrui Liu, Sixiao Zhang, Cheng Long
Furthermore, we introduce a Mixture-of-Experts (MoE) network in each attention head to disentangle various user preferences within each facet.
no code implementations • 26 Oct 2024 • Mingrui Liu, Sixiao Zhang, Cheng Long
If the target document appears in the knowledge database, the masked text will retrieve the complete target document as context, allowing for accurate mask prediction.
1 code implementation • 29 Sep 2024 • Qianxiong Xu, Xuanyi Liu, Lanyun Zhu, Guosheng Lin, Cheng Long, Ziyue Li, Rui Zhao
Hence, we aim to devise a cross (attention-like) Mamba to capture inter-sequence dependencies for FSS.
Ranked #6 on
Few-Shot Semantic Segmentation
on COCO-20i (5-shot)
1 code implementation • 16 Sep 2024 • Jianyang Gao, Yutong Gou, Yuexuan Xu, Yongyi Yang, Cheng Long, Raymond Chi-Wing Wong
In this paper, we introduce a new quantization method to address this limitation by extending RaBitQ.
1 code implementation • 4 Sep 2024 • Yuexuan Xu, Jianyang Gao, Yutong Gou, Cheng Long, Christian S. Jensen
In this study, instead of materializing a compressed index for every possible query range in preparation for querying, we materialize graph-based indexes, called elemental graphs, for a moderate number of ranges.
no code implementations • 18 Jul 2024 • Qiuyu Zhu, Liang Zhang, Qianxiong Xu, Kaijun Liu, Cheng Long, Xiaoyang Wang
Based on this structure, we propose a novel Hierarchical Heterogeneous Graph Transformer (HHGT) model, which seamlessly integrates a Type-level Transformer for aggregating nodes of different types within each k-ring neighborhood, followed by a Ring-level Transformer for aggregating different k-ring neighborhoods in a hierarchical manner.
2 code implementations • 17 Jul 2024 • Sixiao Zhang, Cheng Long, Wei Yuan, Hongxu Chen, Hongzhi Yin
To assess the efficacy of the watermark, the model is tasked with predicting the subsequent item given a truncated watermark sequence.
1 code implementation • 13 Jul 2024 • Qianxiong Xu, Guosheng Lin, Chen Change Loy, Cheng Long, Ziyue Li, Rui Zhao
Recent advancements in few-shot segmentation (FSS) have exploited pixel-by-pixel matching between query and support features, typically based on cross attention, which selectively activate query foreground (FG) features that correspond to the same-class support FG features.
Ranked #7 on
Few-Shot Semantic Segmentation
on PASCAL-5i (1-Shot)
3 code implementations • 3 Jun 2024 • Chenxi Liu, Qianxiong Xu, Hao Miao, Sun Yang, Lingzheng Zhang, Cheng Long, Ziyue Li, Rui Zhao
As another key design, to reduce the computational costs from time series with their length textual prompts, we design an effective prompt to encourage the most essential temporal information to be encapsulated in the last token: only the last token is passed to downstream prediction.
Ranked #8 on
Time Series Forecasting
on ETTh1 (336) Multivariate
1 code implementation • 21 May 2024 • Jianyang Gao, Cheng Long
Recently, with the help of fast SIMD-based implementations, Product Quantization (PQ) and its variants can often efficiently and accurately estimate the distances between the vectors and have achieved great success in the in-memory ANN search.
no code implementations • 15 May 2024 • Qianru Zhang, Haixin Wang, Cheng Long, Liangcai Su, Xingwei He, Jianlong Chang, Tailin Wu, Hongzhi Yin, Siu-Ming Yiu, Qi Tian, Christian S. Jensen
By integrating generative techniques and providing a standardized framework, the paper contributes to advancing the field and encourages researchers to explore the vast potential of generative techniques in spatial-temporal data mining.
no code implementations • 18 Mar 2024 • Yile Chen, Xiucheng Li, Gao Cong, Zhifeng Bao, Cheng Long
In this study, we introduce a novel framework called Toast for learning general-purpose representations of road networks, along with its advanced counterpart DyToast, designed to enhance the integration of temporal dynamics to boost the performance of various time-sensitive downstream tasks.
no code implementations • International Conference on Image Processing (ICIP) 2022 • Cheng Long, Adrian Barbu
Shape modeling is a challenging task with many potential applications in computer vision and medical imaging.
1 code implementation • 6 Feb 2024 • Kethmi Hirushini Hettige, Jiahao Ji, Shili Xiang, Cheng Long, Gao Cong, Jingyuan Wang
Air quality prediction and modelling plays a pivotal role in public health and environment management, for individuals and authorities to make informed decisions.
2 code implementations • 18 Jan 2024 • Chenxi Liu, Sun Yang, Qianxiong Xu, Zhishuai Li, Cheng Long, Ziyue Li, Rui Zhao
In the ST-LLM, we define timesteps at each location as tokens and design a spatial-temporal embedding to learn the spatial location and global temporal patterns of these tokens.
1 code implementation • 5 Nov 2023 • Qianxiong Xu, Cheng Long, Ziyue Li, Sijie Ruan, Rui Zhao, Zhishuai Li
To address this issue, we first present a novel Increment training strategy: instead of masking nodes (and reconstructing them), we add virtual nodes into the training graph so as to mitigate the graph gap issue naturally.
1 code implementation • 25 Oct 2023 • Sixiao Zhang, Hongzhi Yin, Hongxu Chen, Cheng Long
These gradients are used to compute a swap loss, which maximizes the loss of the student model.
no code implementations • 16 Oct 2023 • Jiahao Ji, Jingyuan Wang, Yu Mou, Cheng Long
The framework consists of two main components: an automatic graph decomposition module that decomposes the original graph structure inherent in ST data into subgraphs corresponding to different factors, and a decomposed learning network that learns the partial ST data on each subgraph separately and integrates them for the final prediction.
1 code implementation • ICCV 2023 • Qianxiong Xu, Wenting Zhao, Guosheng Lin, Cheng Long
Moreover, when calculating SCCA, we design a scaled-cosine mechanism to better utilize the support features for similarity calculation.
Ranked #12 on
Few-Shot Semantic Segmentation
on COCO-20i (5-shot)
no code implementations • 3 Jul 2023 • Xinhang Li, Xiangyu Zhao, Yejing Wang, Yu Liu, Yong Li, Cheng Long, Yong Zhang, Chunxiao Xing
As a representative information retrieval task, site recommendation, which aims at predicting the optimal sites for a brand or an institution to open new branches in an automatic data-driven way, is beneficial and crucial for brand development in modern business.
no code implementations • 13 Apr 2023 • Liang Zhang, Cheng Long
The constructed hypergraph would naturally capture the high-order relationships among roads with hyperedges.
1 code implementation • 22 Mar 2023 • Qianxiong Xu, Cheng Long, Liang Yu, Chen Zhang
In this paper, we propose to conduct road extraction based on satellite images and partial road maps, which is new.
1 code implementation • 17 Mar 2023 • Jianyang Gao, Cheng Long
Approximate K nearest neighbor (AKNN) search is a fundamental and challenging problem.
1 code implementation • 15 Nov 2022 • Liang Zhang, Cheng Long, Gao Cong
Motivated by the success of contrastive learning for representation learning, we propose to leverage it for multi-view region representation learning and design a model called ReMVC (Region Embedding with Multi-View Contrastive Learning) by following two guidelines: i) comparing a region with others within each view for effective representation extraction and ii) comparing a region with itself across different views for cross-view information sharing.
1 code implementation • 15 Nov 2022 • Zheng Wang, Mingrui Liu, Cheng Long, Qianru Zhang, Jiangneng Li, Chunyan Miao
The DeepSEI model incorporates two networks called deep network and recurrent network, which extract the features of the mobility records from three aspects, namely spatiality, temporality and activity, one at a coarse level and the other at a detailed level.
1 code implementation • 13 Nov 2022 • Kaixin Wang, Cheng Long, Da Yan, Jie Zhang, H. V. Jagadish
Specifically, we propose a weighted sampling algorithm called WSD for estimating the subgraph count in a fully dynamic graph stream, which samples the edges based on their weights that indicate their importance and reflect their properties.
1 code implementation • 12 Nov 2022 • Qianru Zhang, Zheng Wang, Cheng Long, Chao Huang, Siu-Ming Yiu, Yiding Liu, Gao Cong, Jieming Shi
Detecting anomalous trajectories has become an important task in many location-based applications.
no code implementations • 24 May 2022 • Shaowen Zhou, Bowen Yu, Aixin Sun, Cheng Long, Jingyang Li, Haiyang Yu, Jian Sun, Yongbin Li
Open Information Extraction (OpenIE) facilitates domain-independent discovery of relational facts from large corpora.
Ranked #1 on
Open Information Extraction
on CaRB
Knowledge Base Construction
Natural Language Understanding
+2
no code implementations • 28 Feb 2022 • Yile Chen, Xiucheng Li, Gao Cong, Cheng Long, Zhifeng Bao, Shang Liu, Wanli Gu, Fuzheng Zhang
As a fundamental component in location-based services, inferring the relationship between points-of-interests (POIs) is very critical for service providers to offer good user experience to business owners and customers.
no code implementations • 8 Mar 2021 • Tu Gu, Kaiyu Feng, Gao Cong, Cheng Long, Zheng Wang, Sheng Wang
Learned indices have been proposed to replace classic index structures like B-Tree with machine learning (ML) models.
no code implementations • 31 Jan 2021 • Pan Xiong, Lei Tong, Kun Zhang, Xuhui Shen, Roberto Battiston, Dimitar Ouzounov, Roberto Iuppa, Danny Crookes, Cheng Long, Huiyu Zhou
Amongst the available technologies for earthquake research, remote sensing has been commonly used due to its unique features such as fast imaging and wide image-acquisition range.
no code implementations • 5 Mar 2020 • Zheng Wang, Cheng Long, Gao Cong, Yiding Liu
Similar trajectory search is a fundamental problem and has been well studied over the past two decades.
no code implementations • 28 Feb 2019 • Ce Ju, Zheng Wang, Cheng Long, Xiao-Yu Zhang, Gao Cong, Dong Eui Chang
Forecasting the motion of surrounding obstacles (vehicles, bicycles, pedestrians and etc.)
Robotics I.2.9; I.2.0
no code implementations • 17 Dec 2018 • Zheng Wang, Ce Ju, Gao Cong, Cheng Long
Recently, the topic of graph representation learning has received plenty of attention.