no code implementations • 20 Nov 2024 • Weicai Ye, Xinyu Chen, Ruohao Zhan, Di Huang, Xiaoshui Huang, Haoyi Zhu, Hujun Bao, Wanli Ouyang, Tong He, Guofeng Zhang
To tackle these challenges, we propose a dynamic-aware tracking any point (DATAP) method that leverages consistent video depth and point tracking.
1 code implementation • 29 Oct 2024 • Lin Li, Xinchun Yu, Xinyu Chen, Peng Liang
Public Code Review (PCR) is an assistant to the internal code review of the development team, in the form of a public Software Question Answering (SQA) community, to help developers access high-quality and efficient review services.
no code implementations • 2 Sep 2024 • Xinyu Chen, HanQin Cai, Fuqiang Liu, Jinhua Zhao
This study addresses the problem of convolutional kernel learning in univariate, multivariate, and multidimensional time series data, which is crucial for interpreting temporal patterns in time series and supporting downstream machine learning tasks.
no code implementations • 17 Jun 2024 • Yunxin Li, Xinyu Chen, Baotian Hu, Longyue Wang, Haoyuan Shi, Min Zhang
Through a comprehensive and quantitative evaluation of cutting-edge models, we reveal that: 1) Video-LMMs face difficulties in fine-grained video tasks involving temporal location, object tracking, and anomaly detection; 2) Video-LMMs present inferior logical and relation reasoning abilities; 3) Open-source Video-LMMs' performance is significantly lower than GPT-4o and Gemini-1. 5, lagging by 20 points.
no code implementations • 11 Apr 2024 • Xinyu Chen, Lin Li, Rui Zhang, Peng Liang
Public Code Review (PCR) can be implemented through a Software Question Answering (SQA) community, which facilitates high knowledge dissemination.
no code implementations • 21 Feb 2024 • Yunxin Li, Xinyu Chen, Baotain Hu, Min Zhang
Long video understanding is a significant and ongoing challenge in the intersection of multimedia and artificial intelligence.
1 code implementation • 21 Feb 2024 • Yunxin Li, Xinyu Chen, Baotian Hu, Haoyuan Shi, Min Zhang
Evaluating and Rethinking the current landscape of Large Multimodal Models (LMMs), we observe that widely-used visual-language projection approaches (e. g., Q-former or MLP) focus on the alignment of image-text descriptions yet ignore the visual knowledge-dimension alignment, i. e., connecting visuals to their relevant knowledge.
1 code implementation • 13 Nov 2023 • Yunxin Li, Longyue Wang, Baotian Hu, Xinyu Chen, Wanqi Zhong, Chenyang Lyu, Wei Wang, Min Zhang
The emergence of multimodal large models (MLMs) has significantly advanced the field of visual understanding, offering remarkable capabilities in the realm of visual question answering (VQA).
no code implementations • 26 Jul 2023 • Chao Zhang, Xinyu Chen, Wensheng Li, Lixue Liu, Wei Wu, DaCheng Tao
In this paper, we measure the linear separability of hidden layer outputs to study the characteristics of deep neural networks.
no code implementations • 17 May 2023 • Xinyu Chen, Dalei Yu, Xinyu Zhang
The random forest (RF) algorithm has become a very popular prediction method for its great flexibility and promising accuracy.
1 code implementation • 8 May 2023 • Yunxin Li, Baotian Hu, Xinyu Chen, Yuxin Ding, Lin Ma, Min Zhang
This makes the language model well-suitable for such multi-modal reasoning scenario on joint textual and visual clues.
1 code implementation • 5 May 2023 • Yunxin Li, Baotian Hu, Xinyu Chen, Lin Ma, Yong Xu, Min Zhang
LMEye addresses this issue by allowing the LLM to request the desired visual information aligned with various human instructions, which we term as the dynamic visual information interaction.
1 code implementation • 3 Dec 2022 • Xinyu Chen, Zhanhong Cheng, HanQin Cai, Nicolas Saunier, Lijun Sun
In this study, we first introduce a Laplacian kernel to temporal regularization for characterizing local trends in traffic time series, which can be formulated as a circular convolution.
1 code implementation • 28 Nov 2022 • Xinyu Chen, ChengYuan Zhang, Xiaoxu Chen, Nicolas Saunier, Lijun Sun
In the temporal context, the complex time-varying system behaviors can be revealed by the temporal modes in the proposed model.
no code implementations • 21 Nov 2022 • Ting Han, Kunhao Pan, Xinyu Chen, Dingjie Song, Yuchen Fan, Xinyu Gao, Ruyi Gan, Jiaxing Zhang
Bidirectional Encoder Representations from Transformers or BERT~\cite{devlin-etal-2019-bert} has been one of the base models for various NLP tasks due to its remarkable performance.
no code implementations • 17 May 2022 • Xinyu Chen, Renjie Li, Yueyao Yu, Yuanwen Shen, Wenye Li, Zhaoyu Zhang, Yin Zhang
In this work, we propose the first-ever Transformer model (POViT) to efficiently design and simulate semiconductor photonic devices with multiple objectives.
2 code implementations • 11 May 2022 • Yawei Li, Kai Zhang, Radu Timofte, Luc van Gool, Fangyuan Kong, Mingxi Li, Songwei Liu, Zongcai Du, Ding Liu, Chenhui Zhou, Jingyi Chen, Qingrui Han, Zheyuan Li, Yingqi Liu, Xiangyu Chen, Haoming Cai, Yu Qiao, Chao Dong, Long Sun, Jinshan Pan, Yi Zhu, Zhikai Zong, Xiaoxiao Liu, Zheng Hui, Tao Yang, Peiran Ren, Xuansong Xie, Xian-Sheng Hua, Yanbo Wang, Xiaozhong Ji, Chuming Lin, Donghao Luo, Ying Tai, Chengjie Wang, Zhizhong Zhang, Yuan Xie, Shen Cheng, Ziwei Luo, Lei Yu, Zhihong Wen, Qi Wu1, Youwei Li, Haoqiang Fan, Jian Sun, Shuaicheng Liu, Yuanfei Huang, Meiguang Jin, Hua Huang, Jing Liu, Xinjian Zhang, Yan Wang, Lingshun Long, Gen Li, Yuanfan Zhang, Zuowei Cao, Lei Sun, Panaetov Alexander, Yucong Wang, Minjie Cai, Li Wang, Lu Tian, Zheyuan Wang, Hongbing Ma, Jie Liu, Chao Chen, Yidong Cai, Jie Tang, Gangshan Wu, Weiran Wang, Shirui Huang, Honglei Lu, Huan Liu, Keyan Wang, Jun Chen, Shi Chen, Yuchun Miao, Zimo Huang, Lefei Zhang, Mustafa Ayazoğlu, Wei Xiong, Chengyi Xiong, Fei Wang, Hao Li, Ruimian Wen, Zhijing Yang, Wenbin Zou, Weixin Zheng, Tian Ye, Yuncheng Zhang, Xiangzhen Kong, Aditya Arora, Syed Waqas Zamir, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan, Dandan Gaoand Dengwen Zhouand Qian Ning, Jingzhu Tang, Han Huang, YuFei Wang, Zhangheng Peng, Haobo Li, Wenxue Guan, Shenghua Gong, Xin Li, Jun Liu, Wanjun Wang, Dengwen Zhou, Kun Zeng, Hanjiang Lin, Xinyu Chen, Jinsheng Fang
The aim was to design a network for single image super-resolution that achieved improvement of efficiency measured according to several metrics including runtime, parameters, FLOPs, activations, and memory consumption while at least maintaining the PSNR of 29. 00dB on DIV2K validation set.
1 code implementation • 20 Mar 2022 • Xinyu Chen, ChengYuan Zhang, Xi-Le Zhao, Nicolas Saunier, Lijun Sun
Modern time series datasets are often high-dimensional, incomplete/sparse, and nonstationary.
no code implementations • 1 Feb 2022 • Xinyu Chen, Ian Beaver
The clustering quality exceeds the state-of-the-art results when BERT is fine-tuned with labeled subsets of only 2. 5% of the task data.
no code implementations • 1 Feb 2022 • Xinyu Chen, Ian Beaver
Mining the latent intentions from large volumes of natural language inputs is a key step to help data analysts design and refine Intelligent Virtual Assistants (IVAs) for customer service and sales support.
no code implementations • 12 Aug 2021 • Hao Ming, Xinyu Chen, Xiansong Fang, Lei Zhang, Chenjia Li, Fan Zhang
In this paper, we propose a center-oriented long short-term memory network (Co-LSTM) incorporating a simplified mode with a recycling mechanism in the equalization operation, which can mitigate fiber nonlinearity in coherent optical communication systems with ultralow complexity.
1 code implementation • 30 Apr 2021 • Xinyu Chen, MengYing Lei, Nicolas Saunier, Lijun Sun
In this paper, we propose a low-rank autoregressive tensor completion (LATC) framework by introducing \textit{temporal variation} as a new regularization term into the completion of a third-order (sensor $\times$ time of day $\times$ day) tensor.
1 code implementation • 7 Aug 2020 • Xinyu Chen, Yixian Chen, Nicolas Saunier, Lijun Sun
Recent studies based on tensor nuclear norm have demonstrated the superiority of tensor learning in imputation tasks by effectively characterizing the complex correlations/dependencies in spatiotemporal data.
no code implementations • 1 Jul 2020 • Pu Ren, Xinyu Chen, Lijun Sun, Hao Sun
To address this fundamental issue, this paper presents an incremental Bayesian tensor learning method for reconstruction of spatiotemporal missing data in SHM and forecasting of structural response.
1 code implementation • 18 Jun 2020 • Xinyu Chen, Lijun Sun
In this paper, we propose a low-rank autoregressive tensor completion (LATC) framework to model multivariate time series data.
1 code implementation • 23 Mar 2020 • Xinyu Chen, Jinming Yang, Lijun Sun
Sparsity and missing data problems are very common in spatiotemporal traffic data collected from various sensing systems.
3 code implementations • 14 Oct 2019 • Xinyu Chen, Lijun Sun
In this paper, we propose a Bayesian temporal factorization (BTF) framework for modeling multidimensional time series -- in particular spatiotemporal data -- in the presence of missing values.
no code implementations • 26 Feb 2019 • Chuangyi Gui, Long Zheng, Bingsheng He, Cheng Liu, Xinyu Chen, Xiaofei Liao, Hai Jin
Graph is a well known data structure to represent the associated relationships in a variety of applications, e. g., data science and machine learning.
Distributed, Parallel, and Cluster Computing
no code implementations • 16 Jun 2018 • Xinyu Chen, Qiang Guan, Li-Ta Lo, Simon Su, James Ahrens, Trilce Estrada
We present in situ TensorView to visualize the training and functioning of CNNs as if they are systems of scientific simulations.