Search Results for author: Qifan Wang

Found 73 papers, 31 papers with code

Parallel Decoding via Hidden Transfer for Lossless Large Language Model Acceleration

no code implementations18 Apr 2024 Pengfei Wu, Jiahao Liu, Zhuocheng Gong, Qifan Wang, Jinpeng Li, Jingang Wang, Xunliang Cai, Dongyan Zhao

In this paper, we propose a novel parallel decoding approach, namely \textit{hidden transfer}, which decodes multiple successive tokens simultaneously in a single forward pass.

Language Modelling Large Language Model

Q-PEFT: Query-dependent Parameter Efficient Fine-tuning for Text Reranking with Large Language Models

no code implementations6 Apr 2024 Zhiyuan Peng, Xuyang Wu, Qifan Wang, Sravanthi Rajanala, Yi Fang

Parameter Efficient Fine-Tuning (PEFT) methods have been extensively utilized in Large Language Models (LLMs) to improve the down-streaming tasks without the cost of fine-tuing the whole LLMs.

Text Reranking

Think Twice Before Assure: Confidence Estimation for Large Language Models through Reflection on Multiple Answers

no code implementations15 Mar 2024 Moxin Li, Wenjie Wang, Fuli Feng, Fengbin Zhu, Qifan Wang, Tat-Seng Chua

Confidence estimation aiming to evaluate output trustability is crucial for the application of large language models (LLM), especially the black-box ones.

Enhancing Long-Term Recommendation with Bi-level Learnable Large Language Model Planning

no code implementations29 Feb 2024 Wentao Shi, Xiangnan He, Yang Zhang, Chongming Gao, Xinyue Li, Jizhi Zhang, Qifan Wang, Fuli Feng

To achieve this, we propose a Bi-level Learnable LLM Planner framework, which combines macro-learning and micro-learning through a hierarchical mechanism.

Decision Making Language Modelling +2

PDETime: Rethinking Long-Term Multivariate Time Series Forecasting from the perspective of partial differential equations

no code implementations25 Feb 2024 shiyi qi, Zenglin Xu, Yiduo Li, Liangjian Wen, Qingsong Wen, Qifan Wang, Yuan Qi

Recent advancements in deep learning have led to the development of various models for long-term multivariate time-series forecasting (LMTF), many of which have shown promising results.

Multivariate Time Series Forecasting Time Series

Multimodal Instruction Tuning with Conditional Mixture of LoRA

no code implementations24 Feb 2024 Ying Shen, Zhiyang Xu, Qifan Wang, Yu Cheng, Wenpeng Yin, Lifu Huang

Multimodal Large Language Models (MLLMs) have demonstrated remarkable proficiency in diverse tasks across different domains, with an increasing focus on improving their zero-shot generalization capabilities for unseen multimodal tasks.

Zero-shot Generalization

Vision-Flan: Scaling Human-Labeled Tasks in Visual Instruction Tuning

no code implementations18 Feb 2024 Zhiyang Xu, Chao Feng, Rulin Shao, Trevor Ashby, Ying Shen, Di Jin, Yu Cheng, Qifan Wang, Lifu Huang

Despite vision-language models' (VLMs) remarkable capabilities as versatile visual assistants, two substantial challenges persist within the existing VLM frameworks: (1) lacking task diversity in pretraining and visual instruction tuning, and (2) annotation error and bias in GPT-4 synthesized instruction tuning data.

Hallucination Visual Question Answering

Facing the Elephant in the Room: Visual Prompt Tuning or Full Finetuning?

no code implementations23 Jan 2024 Cheng Han, Qifan Wang, Yiming Cui, Wenguan Wang, Lifu Huang, Siyuan Qi, Dongfang Liu

As the scale of vision models continues to grow, the emergence of Visual Prompt Tuning (VPT) as a parameter-efficient transfer learning technique has gained attention due to its superior performance compared to traditional full-finetuning.

Transfer Learning Visual Prompt Tuning

Jack of All Tasks, Master of Many: Designing General-purpose Coarse-to-Fine Vision-Language Model

no code implementations19 Dec 2023 Shraman Pramanick, Guangxing Han, Rui Hou, Sayan Nag, Ser-Nam Lim, Nicolas Ballas, Qifan Wang, Rama Chellappa, Amjad Almahairi

In this work, we introduce VistaLLM, a powerful visual system that addresses coarse- and fine-grained VL tasks over single and multiple input images using a unified framework.

Attribute Language Modelling +1

RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training

1 code implementation7 Dec 2023 Jaehyung Kim, Yuning Mao, Rui Hou, Hanchao Yu, Davis Liang, Pascale Fung, Qifan Wang, Fuli Feng, Lifu Huang, Madian Khabsa

Under a unified evaluation of fine-tuned LMs by incorporating four representative perspectives of model robustness, we demonstrate the effectiveness of RoAST compared to state-of-the-art fine-tuning methods on six different types of LMs, which indicates its usefulness in practice.

Adversarial Robustness

MART: Improving LLM Safety with Multi-round Automatic Red-Teaming

no code implementations13 Nov 2023 Suyu Ge, Chunting Zhou, Rui Hou, Madian Khabsa, Yi-Chia Wang, Qifan Wang, Jiawei Han, Yuning Mao

Specifically, an adversarial LLM and a target LLM interplay with each other in an iterative manner, where the adversarial LLM aims to generate challenging prompts that elicit unsafe responses from the target LLM, while the target LLM is fine-tuned with safety aligned data on these adversarial prompts.

Instruction Following Response Generation

PsyCoT: Psychological Questionnaire as Powerful Chain-of-Thought for Personality Detection

1 code implementation31 Oct 2023 Tao Yang, Tianyuan Shi, Fanqi Wan, Xiaojun Quan, Qifan Wang, Bingzhe Wu, Jiaxiang Wu

Drawing inspiration from Psychological Questionnaires, which are carefully designed by psychologists to evaluate individual personality traits through a series of targeted items, we argue that these items can be regarded as a collection of well-structured chain-of-thought (CoT) processes.

Improving Input-label Mapping with Demonstration Replay for In-context Learning

no code implementations30 Oct 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

The effectiveness of ICL can be attributed to the strong language modeling capabilities of large language models (LLMs), which enable them to learn the mapping between input and labels based on in-context demonstrations.

In-Context Learning Language Modelling

CoLLM: Integrating Collaborative Embeddings into Large Language Models for Recommendation

1 code implementation30 Oct 2023 Yang Zhang, Fuli Feng, Jizhi Zhang, Keqin Bao, Qifan Wang, Xiangnan He

In pursuit of superior recommendations for both cold and warm start scenarios, we introduce CoLLM, an innovative LLMRec methodology that seamlessly incorporates collaborative information into LLMs for recommendation.

Retrieval-based Knowledge Transfer: An Effective Approach for Extreme Large Language Model Compression

no code implementations24 Oct 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Ran Lucien Wang, Rui Yan

In particular, our approach extracts knowledge from LLMs to construct a knowledge store, from which the small-scale model can retrieve relevant information and leverage it for effective inference.

Language Modelling Large Language Model +3

Dual-Feedback Knowledge Retrieval for Task-Oriented Dialogue Systems

no code implementations23 Oct 2023 Tianyuan Shi, Liangzhi Li, Zijian Lin, Tao Yang, Xiaojun Quan, Qifan Wang

Efficient knowledge retrieval plays a pivotal role in ensuring the success of end-to-end task-oriented dialogue systems by facilitating the selection of relevant information necessary to fulfill user requests.

Open-Domain Question Answering Response Generation +2

Attack Prompt Generation for Red Teaming and Defending Large Language Models

1 code implementation19 Oct 2023 Boyi Deng, Wenjie Wang, Fuli Feng, Yang Deng, Qifan Wang, Xiangnan He

Furthermore, we propose a defense framework that fine-tunes victim LLMs through iterative interactions with the attack framework to enhance their safety against red teaming attacks.

In-Context Learning

Advocating for the Silent: Enhancing Federated Generalization for Non-Participating Clients

no code implementations11 Oct 2023 Zheshun Wu, Zenglin Xu, Dun Zeng, Qifan Wang, Jie Liu

Federated Learning (FL) has surged in prominence due to its capability of collaborative model training without direct data sharing.

Federated Learning Generalization Bounds

Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models

no code implementations7 Oct 2023 Song Jiang, Zahra Shakeri, Aaron Chan, Maziar Sanjabi, Hamed Firooz, Yinglong Xia, Bugra Akyildiz, Yizhou Sun, Jinchao Li, Qifan Wang, Asli Celikyilmaz

Breakdown analysis further highlights RESPROMPT particularly excels in complex multi-step reasoning: for questions demanding at least five reasoning steps, RESPROMPT outperforms the best CoT based benchmarks by a remarkable average improvement of 21. 1% on LLaMA-65B and 14. 3% on LLaMA2-70B.

Math

Enhanced Federated Optimization: Adaptive Unbiased Sampling with Reduced Variance

no code implementations4 Oct 2023 Dun Zeng, Zenglin Xu, Yu Pan, Xu Luo, Qifan Wang, Xiaoying Tang

Central to this process is the technique of unbiased client sampling, which ensures a representative selection of clients.

Federated Learning

Tackling Hybrid Heterogeneity on Federated Optimization via Gradient Diversity Maximization

1 code implementation4 Oct 2023 Dun Zeng, Zenglin Xu, Yu Pan, Qifan Wang, Xiaoying Tang

The combined effects of statistical and system heterogeneity can significantly reduce the efficiency of federated optimization.

Federated Learning

On the Equivalence of Graph Convolution and Mixup

no code implementations29 Sep 2023 Xiaotian Han, Hanqing Zeng, Yu Chen, Shaoliang Nie, Jingzhou Liu, Kanika Narang, Zahra Shakeri, Karthik Abinav Sankararaman, Song Jiang, Madian Khabsa, Qifan Wang, Xia Hu

We establish this equivalence mathematically by demonstrating that graph convolution networks (GCN) and simplified graph convolution (SGC) can be expressed as a form of Mixup.

Data Augmentation

ClusterFormer: Clustering As A Universal Visual Learner

1 code implementation22 Sep 2023 James C. Liang, Yiming Cui, Qifan Wang, Tong Geng, Wenguan Wang, Dongfang Liu

This paper presents CLUSTERFORMER, a universal vision model that is based on the CLUSTERing paradigm with TransFORMER.

Clustering Image Classification +7

LM-Infinite: Zero-Shot Extreme Length Generalization for Large Language Models

1 code implementation30 Aug 2023 Chi Han, Qifan Wang, Hao Peng, Wenhan Xiong, Yu Chen, Heng Ji, Sinong Wang

As a result, their performance suffers drastically on inputs longer than those encountered during training, substantially limiting their applications in real-world tasks involving long contexts such as encoding scientific articles, code repositories, or long dialogues.

2k 4k +1

mCL-NER: Cross-Lingual Named Entity Recognition via Multi-view Contrastive Learning

no code implementations17 Aug 2023 Ying Mo, Jian Yang, Jiahao Liu, Qifan Wang, Ruoyu Chen, Jingang Wang, Zhoujun Li

A multi-view contrastive learning framework is introduced to encompass semantic contrasts between source, codeswitched, and target sentences, as well as contrasts among token-to-token relations.

Contrastive Learning named-entity-recognition +2

E^2VPT: An Effective and Efficient Approach for Visual Prompt Tuning

1 code implementation ICCV 2023 Cheng Han, Qifan Wang, Yiming Cui, Zhiwen Cao, Wenguan Wang, Siyuan Qi, Dongfang Liu

Specifically, we introduce a set of learnable key-value prompts and visual prompts into self-attention and input layers, respectively, to improve the effectiveness of model fine-tuning.

Visual Prompt Tuning

LLM-Rec: Personalized Recommendation via Prompting Large Language Models

no code implementations24 Jul 2023 Hanjia Lyu, Song Jiang, Hanqing Zeng, Yinglong Xia, Qifan Wang, Si Zhang, Ren Chen, Christopher Leung, Jiajie Tang, Jiebo Luo

Notably, the success of LLM-Rec lies in its prompting strategies, which effectively tap into the language model's comprehension of both general and specific item characteristics.

Soft Prompt Tuning for Augmenting Dense Retrieval with Large Language Models

1 code implementation17 Jul 2023 Zhiyuan Peng, Xuyang Wu, Qifan Wang, Yi Fang

We design a filter to select high-quality example document-query pairs in the prompt to further improve the quality of weak tagged queries.

Retrieval TAG +1

Recommendation Unlearning via Influence Function

no code implementations5 Jul 2023 Yang Zhang, Zhiyu Hu, Yimeng Bai, Fuli Feng, Jiancan Wu, Qifan Wang, Xiangnan He

In this work, we propose an Influence Function-based Recommendation Unlearning (IFRU) framework, which efficiently updates the model without retraining by estimating the influence of the unusable data on the model via the influence function.

Meta-training with Demonstration Retrieval for Efficient Few-shot Learning

no code implementations30 Jun 2023 Aaron Mueller, Kanika Narang, Lambert Mathias, Qifan Wang, Hamed Firooz

Meta-training allows one to leverage smaller models for few-shot generalization in a domain-general and task-agnostic manner; however, these methods alone results in models that may not have sufficient parameterization or knowledge to adapt quickly to a large variety of tasks.

Few-Shot Learning QNLI +3

FFB: A Fair Fairness Benchmark for In-Processing Group Fairness Methods

1 code implementation15 Jun 2023 Xiaotian Han, Jianfeng Chi, Yu Chen, Qifan Wang, Han Zhao, Na Zou, Xia Hu

This paper introduces the Fair Fairness Benchmark (\textsf{FFB}), a benchmarking framework for in-processing group fairness methods.

Benchmarking Fairness

PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models

no code implementations30 May 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Yang Yang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Rui Yan

While transformer-based pre-trained language models (PLMs) have dominated a number of NLP applications, these models are heavy to deploy and expensive to use.

Quantization

RankCSE: Unsupervised Sentence Representations Learning via Learning to Rank

1 code implementation26 May 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Kai Chen, Rui Yan

In this paper, we propose a novel approach, RankCSE, for unsupervised sentence representation learning, which incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.

Contrastive Learning Learning-To-Rank +4

AMELI: Enhancing Multimodal Entity Linking with Fine-Grained Attributes

no code implementations24 May 2023 Barry Menglong Yao, Yu Chen, Qifan Wang, Sijia Wang, Minqian Liu, Zhiyang Xu, Licheng Yu, Lifu Huang

We propose attribute-aware multimodal entity linking, where the input is a mention described with a text and image, and the goal is to predict the corresponding target entity from a multimodal knowledge base (KB) where each entity is also described with a text description, a visual image and a set of attributes and values.

Attribute Entity Linking

The Art of SOCRATIC QUESTIONING: Recursive Thinking with Large Language Models

1 code implementation24 May 2023 Jingyuan Qi, Zhiyang Xu, Ying Shen, Minqian Liu, Di Jin, Qifan Wang, Lifu Huang

Chain-of-Thought (CoT) prompting enables large language models to solve complex reasoning problems by generating intermediate steps.

Language Modelling Math +2

Disentangled Phonetic Representation for Chinese Spelling Correction

1 code implementation24 May 2023 Zihong Liang, Xiaojun Quan, Qifan Wang

Chinese Spelling Correction (CSC) aims to detect and correct erroneous characters in Chinese texts.

Spelling Correction

Coarse-to-Fine Contrastive Learning in Image-Text-Graph Space for Improved Vision-Language Compositionality

no code implementations23 May 2023 Harman Singh, Pengchuan Zhang, Qifan Wang, Mengjiao Wang, Wenhan Xiong, Jingfei Du, Yu Chen

Along with this, we propose novel negative mining techniques in the scene graph space for improving attribute binding and relation understanding.

 Ranked #1 on Image Retrieval on CREPE (Compositional REPresentation Evaluation) (Recall@1 (HN-Comp, UC) metric)

Attribute Contrastive Learning +4

AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression

1 code implementation17 May 2023 Siyue Wu, Hongzhan Chen, Xiaojun Quan, Qifan Wang, Rui Wang

To enhance the knowledge transfer of model reasoning and generalization, we further explore multi-view attribution distillation on all potential decisions of the teacher.

Knowledge Distillation Language Modelling +2

Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales

1 code implementation11 May 2023 Brihi Joshi, Ziyi Liu, Sahana Ramnath, Aaron Chan, Zhewei Tong, Shaoliang Nie, Qifan Wang, Yejin Choi, Xiang Ren

Existing metrics like task performance of the LM generating the rationales, or similarity between generated and gold rationales are not good indicators of their human utility.

MMViT: Multiscale Multiview Vision Transformers

no code implementations28 Apr 2023 Yuchen Liu, Natasha Ong, Kaiyan Peng, Bo Xiong, Qifan Wang, Rui Hou, Madian Khabsa, Kaiyue Yang, David Liu, Donald S. Williamson, Hanchao Yu

Our model encodes different views of the input signal and builds several channel-resolution feature stages to process the multiple views of the input at different resolutions in parallel.

Image Classification

Prediction then Correction: An Abductive Prediction Correction Method for Sequential Recommendation

1 code implementation27 Apr 2023 Yulong Huang, Yang Zhang, Qifan Wang, Chenxu Wang, Fuli Feng

To improve the accuracy of these models, some researchers have attempted to simulate human analogical reasoning to correct predictions for testing data by drawing analogies with the prediction errors of similar training data.

Sequential Recommendation

TransFlow: Transformer as Flow Learner

no code implementations CVPR 2023 Yawen Lu, Qifan Wang, Siqi Ma, Tong Geng, Yingjie Victor Chen, Huaijin Chen, Dongfang Liu

Optical flow is an indispensable building block for various important computer vision tasks, including motion estimation, object tracking, and disparity measurement.

Motion Estimation object-detection +4

Defending Against Patch-based Backdoor Attacks on Self-Supervised Learning

1 code implementation CVPR 2023 Ajinkya Tejankar, Maziar Sanjabi, Qifan Wang, Sinong Wang, Hamed Firooz, Hamed Pirsiavash, Liang Tan

It was shown that an adversary can poison a small part of the unlabeled data so that when a victim trains an SSL model on it, the final model will have a backdoor that the adversary can exploit.

Data Poisoning Self-Supervised Learning

SVT: Supertoken Video Transformer for Efficient Video Understanding

no code implementations1 Apr 2023 Chenbin Pan, Rui Hou, Hanchao Yu, Qifan Wang, Senem Velipasalar, Madian Khabsa

Whether by processing videos with fixed resolution from start to end or incorporating pooling and down-scaling strategies, existing video transformers process the whole video content throughout the network without specially handling the large portions of redundant information.

Video Understanding

Stochastic Clustered Federated Learning

no code implementations2 Mar 2023 Dun Zeng, Xiangjing Hu, Shiyu Liu, Yue Yu, Qifan Wang, Zenglin Xu

Federated learning is a distributed learning framework that takes full advantage of private data samples kept on edge devices.

Federated Learning

Representation Deficiency in Masked Language Modeling

1 code implementation4 Feb 2023 Yu Meng, Jitin Krishnan, Sinong Wang, Qifan Wang, Yuning Mao, Han Fang, Marjan Ghazvininejad, Jiawei Han, Luke Zettlemoyer

In this work, we offer a new perspective on the consequence of such a discrepancy: We demonstrate empirically and theoretically that MLM pretraining allocates some model dimensions exclusively for representing $\texttt{[MASK]}$ tokens, resulting in a representation deficiency for real tokens and limiting the pretrained model's expressiveness when it is adapted to downstream data without $\texttt{[MASK]}$ tokens.

Language Modelling Masked Language Modeling

Retiring $Δ$DP: New Distribution-Level Metrics for Demographic Parity

1 code implementation31 Jan 2023 Xiaotian Han, Zhimeng Jiang, Hongye Jin, Zirui Liu, Na Zou, Qifan Wang, Xia Hu

Unfortunately, in this paper, we reveal that the fairness metric $\Delta DP$ can not precisely measure the violation of demographic parity, because it inherently has the following drawbacks: i) zero-value $\Delta DP$ does not guarantee zero violation of demographic parity, ii) $\Delta DP$ values can vary with different classification thresholds.

Fairness

Solve the Puzzle of Instance Segmentation in Videos: A Weakly Supervised Framework with Spatio-Temporal Collaboration

no code implementations15 Dec 2022 Liqi Yan, Qifan Wang, Siqi Ma, Jingang Wang, Changbin Yu

Instance segmentation in videos, which aims to segment and track multiple objects in video frames, has garnered a flurry of research attention in recent years.

Depth Estimation Instance Segmentation +3

Orders Are Unwanted: Dynamic Deep Graph Convolutional Network for Personality Detection

1 code implementation3 Dec 2022 Tao Yang, Jinghao Deng, Xiaojun Quan, Qifan Wang

Predicting personality traits based on online posts has emerged as an important task in many fields such as social network analysis.

Improved Adaptive Algorithm for Scalable Active Learning with Weak Labeler

no code implementations4 Nov 2022 Yifang Chen, Karthik Sankararaman, Alessandro Lazaric, Matteo Pirotta, Dmytro Karamshuk, Qifan Wang, Karishma Mandyam, Sinong Wang, Han Fang

We design a novel algorithmic template, Weak Labeler Active Cover (WL-AC), that is able to robustly leverage the lower quality weak labelers to reduce the query complexity while retaining the desired level of accuracy.

Active Learning

COFFEE: Counterfactual Fairness for Personalized Text Generation in Explainable Recommendation

no code implementations14 Oct 2022 Nan Wang, Qifan Wang, Yi-Chia Wang, Maziar Sanjabi, Jingzhou Liu, Hamed Firooz, Hongning Wang, Shaoliang Nie

However, the bias inherent in user written text, often used for PTG model training, can inadvertently associate different levels of linguistic quality with users' protected attributes.

counterfactual Counterfactual Inference +4

AD-DROP: Attribution-Driven Dropout for Robust Language Model Fine-Tuning

1 code implementation12 Oct 2022 Tao Yang, Jinghao Deng, Xiaojun Quan, Qifan Wang, Shaoliang Nie

Fine-tuning large pre-trained language models on downstream tasks is apt to suffer from overfitting when limited training data is available.

Language Modelling

Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair Modeling

1 code implementation11 Oct 2022 Yuanhang Yang, shiyi qi, Chuanyi Liu, Qifan Wang, Cuiyun Gao, Zenglin Xu

Transformer-based models have achieved great success on sentence pair modeling tasks, such as answer selection and natural language inference (NLI).

Answer Selection Natural Language Inference +2

XPrompt: Exploring the Extreme of Prompt Tuning

no code implementations10 Oct 2022 Fang Ma, Chen Zhang, Lei Ren, Jingang Wang, Qifan Wang, Wei Wu, Xiaojun Quan, Dawei Song

Prompt tuning learns soft prompts to condition frozen Pre-trained Language Models (PLMs) for performing downstream tasks in a parameter-efficient manner.

Rethinking Missing Data: Aleatoric Uncertainty-Aware Recommendation

1 code implementation22 Sep 2022 Chenxu Wang, Fuli Feng, Yang Zhang, Qifan Wang, Xunhan Hu, Xiangnan He

A standard choice is treating the missing data as negative training samples and estimating interaction likelihood between user-item pairs along with the observed interactions.

Fall Detection from Audios with Audio Transformers

no code implementations23 Aug 2022 Prabhjot Kaur, Qifan Wang, Weisong Shi

Our paper provides a novel, non-wearable, non-intrusive, and scalable solution for fall detection, deployed on an autonomous mobile robot equipped with a microphone.

Towards Unbiased Label Distribution Learning for Facial Pose Estimation Using Anisotropic Spherical Gaussian

no code implementations19 Aug 2022 Zhiwen Cao, Dongfang Liu, Qifan Wang, Yingjie Chen

In this paper, we propose an Anisotropic Spherical Gaussian (ASG)-based LDL approach for facial pose estimation.

Pose Estimation

MiniDisc: Minimal Distillation Schedule for Language Model Compression

1 code implementation29 May 2022 Chen Zhang, Yang Yang, Qifan Wang, Jiahao Liu, Jingang Wang, Wei Wu, Dawei Song

In particular, motivated by the finding that the performance of the student is positively correlated to the scale-performance tradeoff of the teacher assistant, MiniDisc is designed with a $\lambda$-tradeoff to measure the optimality of the teacher assistant without trial distillation to the student.

Knowledge Distillation Language Modelling +2

GL-RG: Global-Local Representation Granularity for Video Captioning

1 code implementation22 May 2022 Liqi Yan, Qifan Wang, Yiming Cui, Fuli Feng, Xiaojun Quan, Xiangyu Zhang, Dongfang Liu

Video captioning is a challenging task as it needs to accurately transform visual understanding into natural language description.

Caption Generation Descriptive +1

Deep Partial Multiplex Network Embedding

no code implementations5 Mar 2022 Qifan Wang, Yi Fang, Anirudh Ravula, Ruining He, Bin Shen, Jingang Wang, Xiaojun Quan, Dongfang Liu

Network embedding is an effective technique to learn the low-dimensional representations of nodes in networks.

Link Prediction Network Embedding +1

WebFormer: The Web-page Transformer for Structure Information Extraction

no code implementations1 Feb 2022 Qifan Wang, Yi Fang, Anirudh Ravula, Fuli Feng, Xiaojun Quan, Dongfang Liu

Structure information extraction refers to the task of extracting structured text fields from web pages, such as extracting a product offer from a shopping page including product title, description, brand and price.

Deep Attention document understanding +1

MAVE: A Product Dataset for Multi-source Attribute Value Extraction

1 code implementation16 Dec 2021 Li Yang, Qifan Wang, Zac Yu, Anand Kulkarni, Sumit Sanghai, Bin Shu, Jon Elsas, Bhargav Kanagal

Attribute value extraction refers to the task of identifying values of an attribute of interest from product information.

Attribute Attribute Extraction +2

Should Graph Convolution Trust Neighbors? A Simple Causal Inference Method

1 code implementation22 Oct 2020 Fuli Feng, Weiran Huang, Xiangnan He, Xin Xin, Qifan Wang, Tat-Seng Chua

To this end, we analyze the working mechanism of GCN with causal graph, estimating the causal effect of a node's local structure for the prediction.

Blocking Causal Inference +4

CatGCN: Graph Convolutional Networks with Categorical Node Features

1 code implementation11 Sep 2020 Weijian Chen, Fuli Feng, Qifan Wang, Xiangnan He, Chonggang Song, Guohui Ling, Yongdong Zhang

In this paper, we propose a new GCN model named CatGCN, which is tailored for graph learning when the node features are categorical.

Graph Learning Node Classification +1

Graph Convolution Machine for Context-aware Recommender System

1 code implementation30 Jan 2020 Jiancan Wu, Xiangnan He, Xiang Wang, Qifan Wang, Weijian Chen, Jianxun Lian, Xing Xie

The encoder projects users, items, and contexts into embedding vectors, which are passed to the GC layers that refine user and item embeddings with context-aware graph convolutions on user-item graph.

Collaborative Filtering Recommendation Systems

Ranking Preserving Hashing for Fast Similarity Search

no code implementations AAAI 2015 Qifan Wang, Zhiwei Zhang, Luo Si

But in many real world applications, ranking measure is important for evaluating the quality of hashing codes. In this paper, we propose a novel Ranking Preserving Hashing (RPH) approach that directly optimizes a popular ranking measure, Normalized Discounted Cumulative Gain (NDCG), to obtain effective hashing codes with high ranking accuracy.

Computational Efficiency

Cannot find the paper you are looking for? You can Submit a new open access paper.