1 code implementation • 19 Aug 2024 • Yuran Xiang, Haiteng Zhao, Chang Ma, Zhi-Hong Deng
Recent advancements in computational chemistry have increasingly focused on synthesizing molecules based on textual instructions.
1 code implementation • 13 Aug 2024 • Shibo Jie, Yehui Tang, Jianyuan Guo, Zhi-Hong Deng, Kai Han, Yunhe Wang
Token compression expedites the training and inference of Vision Transformers (ViTs) by reducing the number of the redundant tokens, e. g., pruning inattentive tokens or merging similar tokens.
1 code implementation • 9 May 2024 • Shibo Jie, Yehui Tang, Ning Ding, Zhi-Hong Deng, Kai Han, Yunhe Wang
Current solutions for efficiently constructing large vision-language (VL) models follow a two-step paradigm: projecting the output of pre-trained vision encoders to the input space of pre-trained language models as visual prompts; and then transferring the models to downstream VL tasks via end-to-end parameter-efficient fine-tuning (PEFT).
1 code implementation • 24 Feb 2024 • Haiteng Zhao, Chang Ma, Guoyin Wang, Jing Su, Lingpeng Kong, Jingjing Xu, Zhi-Hong Deng, Hongxia Yang
Large Language Model (LLM) Agents have recently garnered increasing interest yet they are limited in their ability to learn from trial and error, a key element of intelligent behavior.
1 code implementation • ICCV 2023 • Shibo Jie, Haoqing Wang, Zhi-Hong Deng
Current state-of-the-art results in computer vision depend in part on fine-tuning large pre-trained vision models.
1 code implementation • NeurIPS 2023 • Haiteng Zhao, Shengchao Liu, Chang Ma, Hannan Xu, Jie Fu, Zhi-Hong Deng, Lingpeng Kong, Qi Liu
We pretrain GIMLET on the molecule tasks along with instructions, enabling the model to transfer effectively to a broad range of tasks.
1 code implementation • 16 May 2023 • Ziheng Li, Shaohan Huang, Zihan Zhang, Zhi-Hong Deng, Qiang Lou, Haizhen Huang, Jian Jiao, Furu Wei, Weiwei Deng, Qi Zhang
Recent studies have shown that dual encoder models trained with the sentence-level translation ranking task are effective methods for cross-lingual sentence embedding.
1 code implementation • CVPR 2023 • Haoqing Wang, Yehui Tang, Yunhe Wang, Jianyuan Guo, Zhi-Hong Deng, Kai Han
The lower layers are not explicitly guided and the interaction among their patches is only used for calculating new activations.
1 code implementation • 1 Mar 2023 • Haiteng Zhao, Shuming Ma, Dongdong Zhang, Zhi-Hong Deng, Furu Wei
Despite that going deep has proven successful in many neural architectures, the existing graph transformers are relatively shallow.
no code implementations • 23 Feb 2023 • Ziheng Li, Shibo Jie, Zhi-Hong Deng
In continual learning, model needs to continually learn a feature extractor and classifier on a sequence of tasks.
1 code implementation • 6 Dec 2022 • Shibo Jie, Zhi-Hong Deng
Recent work has explored the potential to adapt a pre-trained vision transformer (ViT) by updating only a few parameters so as to improve storage efficiency, called parameter-efficient transfer learning (PETL).
1 code implementation • European Conference on Computer Vision 2022 • Haoqing Wang, Zhi-Hong Deng
This results in our CPN (Contrastive Prototypical Network) model, which combines the prototypical loss with pairwise contrast and outperforms the existing models from this paradigm with modestly large batch size.
1 code implementation • 14 Jul 2022 • Shibo Jie, Zhi-Hong Deng
The pretrain-then-finetune paradigm has been widely adopted in computer vision.
1 code implementation • 24 May 2022 • Haiteng Zhao, Chang Ma, Xinshuai Dong, Anh Tuan Luu, Zhi-Hong Deng, Hanwang Zhang
Deep learning models have achieved great success in many fields, yet they are vulnerable to adversarial examples.
no code implementations • 19 May 2022 • Gehui Shen, Shibo Jie, Ziheng Li, Zhi-Hong Deng
In our framework, a generative classifier which utilizes replay memory is used for inference, and the training objective is a pair-based metric learning loss which is proven theoretically to optimize the feature space in a generative way.
1 code implementation • 22 Apr 2022 • Shibo Jie, Zhi-Hong Deng, Ziheng Li
We study a practical setting of continual learning: fine-tuning on a pre-trained model continually.
1 code implementation • CVPR 2022 • Haoqing Wang, Xun Guo, Zhi-Hong Deng, Yan Lu
It significantly improves the performance of several classic contrastive learning models in downstream tasks.
no code implementations • 12 Mar 2022 • Zhi-Hong Deng, Chang-Dong Wang, Ling Huang, Jian-Huang Lai, Philip S. Yu
G$^3$SR decomposes the session-based recommendation workflow into two steps.
1 code implementation • 23 Oct 2021 • Haiteng Zhao, Chang Ma, Qinyu Chen, Zhi-Hong Deng
In the framework, a surrogate joint distribution models the underlying joint distribution of the unlabeled target domain.
no code implementations • 29 Sep 2021 • Haoqing Wang, Xun Guo, Zhi-Hong Deng, Yan Lu
Therefore, we assume the task-relevant information that is not shared between views can not be ignored and theoretically prove that the minimal sufficient representation in contrastive learning is not sufficient for the downstream tasks, which causes performance degradation.
1 code implementation • 29 Apr 2021 • Haoqing Wang, Zhi-Hong Deng
However, when there exists the domain shift between the training tasks and the test tasks, the obtained inductive bias fails to generalize across domains, which degrades the performance of the meta-learning models.
Ranked #2 on Cross-Domain Few-Shot on cars
1 code implementation • 10 Mar 2021 • Zi-Yuan Hu, Jin Huang, Zhi-Hong Deng, Chang-Dong Wang, Ling Huang, Jian-Huang Lai, Philip S. Yu
Representation learning tries to learn a common low dimensional space for the representations of users and items.
1 code implementation • 12 Sep 2020 • Haoqing Wang, Zhi-Hong Deng
The performance of meta-learning approaches for few-shot learning generally depends on three aspects: features suitable for comparison, the classifier ( base learner ) suitable for low-data scenarios, and valuable information from the samples to classify.
no code implementations • 10 Jun 2020 • Song Zhang, Gehui Shen, Jinsong Huang, Zhi-Hong Deng
Lifelong or continual learning remains to be a challenge for artificial neural network, as it is required to be both stable for preservation of old knowledge and plastic for acquisition of new knowledge.
no code implementations • 7 May 2020 • Gehui Shen, Song Zhang, Xiang Chen, Zhi-Hong Deng
For this scenario, generative replay is a promising strategy which generates and replays pseudo data for previous tasks to alleviate catastrophic forgetting.
1 code implementation • NeurIPS 2019 • Zhiqing Sun, Zhuohan Li, Haoqing Wang, Zi Lin, Di He, Zhi-Hong Deng
However, these models assume that the decoding process of each token is conditionally independent of others.
1 code implementation • ICLR 2020 • Xiaoran Xu, Wei Feng, Yunsheng Jiang, Xiaohui Xie, Zhiqing Sun, Zhi-Hong Deng
We propose Dynamically Pruned Message Passing Networks (DPMPN) for large-scale knowledge graph reasoning.
1 code implementation • 30 May 2019 • Xiaoran Xu, Wei Feng, Zhiqing Sun, Zhi-Hong Deng
Instead, inspired by the consciousness prior proposed by Yoshua Bengio, we explore reasoning with the notion of attentive awareness from a cognitive perspective, and formulate it in the form of attentive message passing on graphs, called neural consciousness flow (NeuCFlow).
1 code implementation • 28 May 2019 • Ting Huang, Gehui Shen, Zhi-Hong Deng
Compared to previous models which can also skip words, our model achieves better trade-offs between performance and efficiency.
no code implementations • 19 May 2019 • Zhiqing Sun, Jian Tang, Pan Du, Zhi-Hong Deng, Jian-Yun Nie
Furthermore, we propose a diversified point network to generate a set of diverse keyphrases out of the word graph in the decoding process.
10 code implementations • ICLR 2019 • Zhiqing Sun, Zhi-Hong Deng, Jian-Yun Nie, Jian Tang
We study the problem of learning representations of entities and relations in knowledge graphs for predicting missing links.
Ranked #2 on Link Prediction on FB122
2 code implementations • 15 Jan 2019 • Zhi-Hong Deng, Ling Huang, Chang-Dong Wang, Jian-Huang Lai, Philip S. Yu
To solve this problem, many methods have been studied, which can be generally categorized into two types, i. e., representation learning-based CF methods and matching function learning-based CF methods.
1 code implementation • EMNLP 2018 • Zhiqing Sun, Zhi-Hong Deng
As far as we know, we are the first to propose a neural model for unsupervised CWS and achieve competitive performance to the state-of-the-art statistical models on four different datasets from SIGHAN 2005 bakeoff.
no code implementations • 18 Aug 2018 • Gehui Shen, Zhi-Hong Deng, Ting Huang, Xi Chen
Recursive Neural Network (RecNN), a type of models which compose words or phrases recursively over syntactic tree structures, has been proven to have superior ability to obtain sentence representation for a variety of NLP tasks.
no code implementations • COLING 2018 • Meng Zou, Xihan Li, Haokun Liu, Zhi-Hong Deng
Neural encoder-decoder models have been widely applied to conversational response generation, which is a research hot spot in recent years.
no code implementations • 2 Jun 2018 • Xi Chen, Zhi-Hong Deng, Gehui Shen, Ting Huang
This paper proposes a novel framework for recurrent neural networks (RNNs) inspired by the human memory models in the field of cognitive neuroscience to enhance information processing and transmission between adjacent RNNs' units.
no code implementations • 27 Dec 2017 • Zhiqing Sun, Gehui Shen, Zhi-Hong Deng
However, if we consider segmenting a given sentence, the most intuitive idea is to predict whether to segment for each gap between two consecutive characters, which in comparison makes previous approaches seem too complex.
no code implementations • EMNLP 2017 • Gehui Shen, Yunlun Yang, Zhi-Hong Deng
Sentence pair modeling is a crucial problem in the field of natural language processing.
no code implementations • COLING 2016 • Shulei Ma, Zhi-Hong Deng, Yunlun Yang
In the age of information exploding, multi-document summarization is attracting particular attention for the ability to help people get the main ideas in a short time.