no code implementations • 5 Nov 2024 • Shan Zhao, Zhaiyu Chen, Zhitong Xiong, Yilei Shi, Sudipan Saha, Xiao Xiang Zhu
Earth Observation (EO) data analysis has been significantly revolutionized by deep learning (DL), with applications typically limited to grid-like data structures.
no code implementations • 5 Sep 2024 • Md Abu Talha, Yongjia Xu, Shan Zhao, Weihua Geng
By introducing space-depend effects such as diffusion and creation in addition to the SIR model, the Fisher's model is in fact a more advanced and comprehensive model.
no code implementations • 23 May 2024 • Shezheng Song, Shasha Li, Shan Zhao, Chengyu Wang, Xiaopeng Li, Jie Yu, Qian Wan, Jun Ma, Tianwei Yan, Wentao Ma, Xiaoguang Mao
In contrast, a pipeline framework first identifies aspects through MATE (Multimodal Aspect Term Extraction) and then aligns these aspects with image patches for sentiment classification (MASC: Multimodal Aspect-Oriented Sentiment Classification).
Aspect-Based Sentiment Analysis Multimodal Sentiment Analysis +2
1 code implementation • 7 Apr 2024 • Shezheng Song, Shasha Li, Shan Zhao, Xiaopeng Li, Chengyu Wang, Jie Yu, Jun Ma, Tianwei Yan, Bin Ji, Xiaoguang Mao
Multimodal entity linking (MEL) aims to utilize multimodal information (usually textual and visual information) to link ambiguous mentions to unambiguous entities in knowledge base.
no code implementations • 13 Mar 2024 • Shan Zhao, Ioannis Prapas, Ilektra Karasante, Zhitong Xiong, Ioannis Papoutsis, Gustau Camps-Valls, Xiao Xiang Zhu
In that direction, we propose integrating causality with Graph Neural Networks (GNNs) that explicitly model the causal mechanism among complex variables via graph learning.
no code implementations • 31 Jan 2024 • Shan Zhao, Zhitong Xiong, Xiao Xiang Zhu
Subseasonal forecasting, which is pivotal for agriculture, water resource management, and early warning of disasters, faces challenges due to the chaotic nature of the atmosphere.
1 code implementation • 19 Dec 2023 • Shezheng Song, Shan Zhao, Chengyu Wang, Tianwei Yan, Shasha Li, Xiaoguang Mao, Meng Wang
Multimodal Entity Linking (MEL) aims at linking ambiguous mentions with multimodal information to entity in Knowledge Graph (KG) such as Wikipedia, which plays a key role in many applications.
no code implementations • 10 Nov 2023 • Shezheng Song, Xiaopeng Li, Shasha Li, Shan Zhao, Jie Yu, Jun Ma, Xiaoguang Mao, Weimin Zhang
We explore Multimodal Large Language Models (MLLMs), which integrate LLMs like GPT-4 to handle multimodal data, including text, images, audio, and more.
no code implementations • 11 Sep 2023 • Shan Zhao, Sudipan Saha, Zhitong Xiong, Niklas Boers, Xiao Xiang Zhu
Motivated by this, we explore a geometric deep learning-based temporal Graph Convolutional Network (GCN) for precipitation nowcasting.
no code implementations • 16 Mar 2023 • Yukuan Zhang, Yunhua Jia, Housheng Xie, Mengzhen Li, Limin Zhao, Yang Yang, Shan Zhao
However, modeling the motion and appearance models of objects in complex scenes still faces various challenging issues.
1 code implementation • 5 Oct 2022 • Meng Sang, Jiaxuan Chen, Mengzhen Li, Pan Tan, Anning Pan, Shan Zhao, Yang Yang
In the field of face recognition, it is always a hot research topic to improve the loss solution to make the face features extracted by the network have greater discriminative power.
1 code implementation • CVPR 2022 • Fushun Zhu, Shan Zhao, Peng Wang, Hao Wang, Hua Yan, Shuaicheng Liu
We propose a semi-supervised network for wide-angle portraits correction.
no code implementations • 26 Aug 2021 • Sudipan Saha, Shan Zhao, Nasrullah Sheikh, Xiao Xiang Zhu
Multi-target domain adaptation is a powerful extension in which a single classifier is learned for multiple unlabeled target domains.
1 code implementation • CVPR 2021 • Jing Tan, Shan Zhao, Pengfei Xiong, Jiangyu Liu, Haoqiang Fan, Shuaicheng Liu
Wide-angle portraits often enjoy expanded views.
no code implementations • 1 Jul 2020 • Shan Zhao, Minghao Hu, Zhiping Cai, Fang Liu
The network is carefully constructed by stacking multiple attention units in depth to fully model dense interactions over token-label spaces, in which two basic attention units are proposed to explicitly capture fine-grained correlations across different modalities (e. g., token-to-token and labelto-token).
no code implementations • 16 Feb 2020 • Daniel Ranti, Katie Hanss, Shan Zhao, Varun Arvind, Joseph Titano, Anthony Costa, Eric Oermann
The BERT models using either set of pretrained checkpoints outperformed the logistic regression model, achieving sample-weighted average F1-scores of 0. 87 and 0. 87 for the general domain model and the combined general and biomedical-domain model.