no code implementations • 14 Mar 2024 • Yizhe Xiong, Hui Chen, Tianxiang Hao, Zijia Lin, Jungong Han, Yuesong Zhang, Guoxin Wang, Yongjun Bao, Guiguang Ding
Consequently, a simple combination of them cannot guarantee accomplishing both training efficiency and inference efficiency with minimal costs.
1 code implementation • 20 Jan 2024 • Guangyuan Ma, Xing Wu, Zijia Lin, Songlin Hu
In this study, we aim to shed light on this issue by revealing that masked auto-encoder (MAE) pre-training with enhanced decoding significantly improves the term coverage of input tokens in dense representations, compared to vanilla BERT checkpoints.
2 code implementations • 10 Dec 2023 • Ao Wang, Hui Chen, Zijia Lin, Jungong Han, Guiguang Ding
Here, to achieve real-time segmenting anything on mobile devices, following MobileSAM, we replace the heavyweight image encoder in SAM with RepViT model, ending up with the RepViT-SAM model.
1 code implementation • 30 Oct 2023 • Zhenpeng Su, Xing Wu, Xue Bai, Zijia Lin, Hui Chen, Guiguang Ding, Wei Zhou, Songlin Hu
Experiments reveal that models incorporating the proposed MiLe Loss can gain consistent performance improvement on downstream benchmarks.
Ranked #89 on Multi-task Language Understanding on MMLU
no code implementations • 11 Oct 2023 • Jiayi Fu, Lei Lin, Xiaoyang Gao, Pengli Liu, Zhengzong Chen, Zhirui Yang, ShengNan Zhang, Xue Zheng, Yan Li, Yuliang Liu, Xucheng Ye, Yiqiao Liao, Chao Liao, Bin Chen, Chengru Song, Junchen Wan, Zijia Lin, Fuzheng Zhang, Zhongyuan Wang, Di Zhang, Kun Gai
Recent advancements in large language models (LLMs) have demonstrated remarkable abilities in handling a variety of natural language processing (NLP) downstream tasks, even on mathematical tasks requiring multi-step reasoning.
Ranked #87 on Arithmetic Reasoning on GSM8K (using extra training data)
no code implementations • 27 Sep 2023 • Ao Wang, Hui Chen, Zijia Lin, Sicheng Zhao, Jungong Han, Guiguang Ding
We further employ a consistent dynamic channel pruning (CDCP) strategy to dynamically prune unimportant channels in ViTs.
1 code implementation • ICCV 2023 • Yizhe Xiong, Hui Chen, Zijia Lin, Sicheng Zhao, Guiguang Ding
To address this issue, recent works consider the Few-shot Unsupervised Domain Adaptation (FUDA) where only a few source samples are labeled, and conduct knowledge transfer via self-supervised learning methods.
no code implementations • 16 Aug 2023 • Guangyuan Ma, Xing Wu, Peng Wang, Zijia Lin, Songlin Hu
Concretely, we leverage the capabilities of LLMs for document expansion, i. e. query generation, and effectively transfer expanded knowledge to retrievers using pre-training strategies tailored for passage retrieval.
7 code implementations • 18 Jul 2023 • Ao Wang, Hui Chen, Zijia Lin, Jungong Han, Guiguang Ding
Recently, lightweight Vision Transformers (ViTs) demonstrate superior performance and lower latency, compared with lightweight Convolutional Neural Networks (CNNs), on resource-constrained mobile devices.
no code implementations • 5 Apr 2023 • Xing Wu, Guangyuan Ma, Peng Wang, Meng Lin, Zijia Lin, Fuzheng Zhang, Songlin Hu
As an effective representation bottleneck pretraining technique, the contextual masked auto-encoder utilizes contextual embedding to assist in the reconstruction of passages.
2 code implementations • 19 Dec 2022 • Xing Wu, Guangyuan Ma, Wanhui Qian, Zijia Lin, Songlin Hu
Recently, methods have been developed to improve the performance of dense passage retrieval by using context-supervised pre-training.
1 code implementation • 13 Oct 2022 • Xing Wu, Chaochen Gao, Zijia Lin, Zhongyuan Wang, Jizhong Han, Songlin Hu
Sparse sampling is also likely to miss important frames corresponding to some text portions, resulting in textual redundancy.
2 code implementations • 8 Oct 2022 • Xing Wu, Chaochen Gao, Zijia Lin, Jizhong Han, Zhongyuan Wang, Songlin Hu
Contrastive learning has been extensively studied in sentence embedding learning, which assumes that the embeddings of different views of the same sentence are closer.
2 code implementations • 16 Aug 2022 • Xing Wu, Guangyuan Ma, Meng Lin, Zijia Lin, Zhongyuan Wang, Songlin Hu
Dense passage retrieval aims to retrieve the relevant passages of a query from a large corpus based on dense representations (i. e., vectors) of the query and the passages.
1 code implementation • 15 Jul 2020 • Qianhui Wu, Zijia Lin, Börje F. Karlsson, Biqing Huang, Jian-Guang Lou
Prior works in cross-lingual named entity recognition (NER) with no/little labeled data fall into two primary categories: model transfer based and data transfer based methods.
Ranked #1 on Cross-Lingual NER on NoDaLiDa Norwegian Bokmål
no code implementations • 17 Jun 2020 • Yunqi Miao, Zijia Lin, Guiguang Ding, Jungong Han
In this paper, we propose a Shallow feature based Dense Attention Network (SDANet) for crowd counting from still images, which diminishes the impact of backgrounds via involving a shallow feature based attention model, and meanwhile, captures multi-scale information via densely connecting hierarchical image features.
1 code implementation • ACL 2020 • Qianhui Wu, Zijia Lin, Börje F. Karlsson, Jian-Guang Lou, Biqing Huang
However, such methods either are not applicable if the labeled data in the source languages is unavailable, or do not leverage information contained in unlabeled data in the target language.
Ranked #1 on Cross-Lingual NER on CoNLL German
1 code implementation • CVPR 2020 • Hui Chen, Guiguang Ding, Xudong Liu, Zijia Lin, Ji Liu, Jungong Han
Existing methods leverage the attention mechanism to explore such correspondence in a fine-grained manner.
Ranked #18 on Cross-Modal Retrieval on Flickr30k
1 code implementation • 14 Nov 2019 • Qianhui Wu, Zijia Lin, Guoxin Wang, Hui Chen, Börje F. Karlsson, Biqing Huang, Chin-Yew Lin
For languages with no annotated resources, transferring knowledge from rich-resource languages is an effective solution for named entity recognition (NER).
Ranked #1 on Cross-Lingual NER on MSRA
1 code implementation • 12 Jul 2019 • Hui Chen, Zijia Lin, Guiguang Ding, JianGuang Lou, Yusen Zhang, Borje Karlsson
The dominant approaches for named entity recognition (NER) mostly adopt complex recurrent neural networks (RNN), e. g., long-short-term-memory (LSTM).
Ranked #23 on Named Entity Recognition (NER) on Ontonotes v5 (English)
no code implementations • CVPR 2015 • Zijia Lin, Guiguang Ding, Mingqing Hu, Jian-Min Wang
With benefits of low storage costs and high query speeds, hashing methods are widely researched for efficiently retrieving large-scale data, which commonly contains multiple views, e. g. a news report with images, videos and texts.
no code implementations • CVPR 2013 • Zijia Lin, Guiguang Ding, Mingqing Hu, Jian-Min Wang, Xiaojun Ye
Though widely utilized for facilitating image management, user-provided image tags are usually incomplete and insufficient to describe the whole semantic content of corresponding images, resulting in performance degradations in tag-dependent applications and thus necessitating effective tag completion methods.