1 code implementation • 25 Mar 2024 • Han Wang, Yanjie Wang, YongJie Ye, Yuxiang Nie, Can Huang
Multi-modal Large Language Models (MLLMs) have demonstrated their ability to perceive objects in still images, but their application in video-related tasks, such as object tracking, remains understudied.
Ranked #1 on Zero-Shot Single Object Tracking on LaSOT
1 code implementation • 7 Feb 2024 • Jinghui Lu, Ziwei Yang, Yanjie Wang, Xuejing Liu, Brian Mac Namee, Can Huang
In this study, we aim to reduce generation latency for Named Entity Recognition (NER) with Large Language Models (LLMs).
1 code implementation • 8 Jan 2024 • Han Wang, Yanjie Wang, Yang Li, Can Huang
In this paper, we propose a novel Global Video Text Spotting Transformer GloTSFormer to model the tracking problem as global associations and utilize the Gaussian Wasserstein distance to guide the morphological correlation between frames.
no code implementations • 29 Aug 2022 • Zhitong Lai, Haichao Sun, Rui Tian, Nannan Ding, Zhiguo Wu, Yanjie Wang
Skip connections are fundamental units in encoder-decoder networks, which are able to improve the feature propagtion of the neural networks.
no code implementations • 1 Dec 2021 • Yanjie Wang, Xu Zou, Zhijun Zhang, Wenhui Xu, Liqun Chen, Sheng Zhong, Luxin Yan, Guodong Wang
Detecting oriented objects along with estimating their rotation information is one crucial step for analyzing remote sensing images.
1 code implementation • ACL 2020 • Samuel Broscheit, Kiril Gashteovski, Yanjie Wang, Rainer Gemulla
An evaluation in such a setup raises the question if a correct prediction is actually a new fact that was induced by reasoning over the open knowledge graph or if it can be trivially explained.
no code implementations • 3 Feb 2019 • Yanjie Wang, Samuel Broscheit, Rainer Gemulla
We propose the Relational Tucker3 (RT) decomposition for multi-relational link prediction in knowledge graphs.
no code implementations • WS 2019 • Yanjie Wang, Daniel Ruffinelli, Rainer Gemulla, Samuel Broscheit, Christian Meilicke
In this paper, we explore whether recent models work well for knowledge base completion and argue that the current evaluation protocols are more suited for question answering rather than knowledge base completion.
no code implementations • 14 Sep 2017 • Yanjie Wang, Rainer Gemulla, Hui Li
Bilinear models belong to the most basic models for this task, they are comparably efficient to train and use, and they can provide good prediction performance.