Search Results for author: Tao Shen

Found 71 papers, 35 papers with code

Reciprocal Learning of Knowledge Retriever and Response Ranker for Knowledge-Grounded Conversations

no code implementations COLING 2022 Jiazhan Feng, Chongyang Tao, Zhen Li, Chang Liu, Tao Shen, Dongyan Zhao

In this paper, we propose a reciprocal learning approach to jointly optimize a knowledge retriever and a response ranker for knowledge-grounded response retrieval without ground-truth knowledge labels.

Retrieval

Social Norms-Grounded Machine Ethics in Complex Narrative Situation

no code implementations COLING 2022 Tao Shen, Xiubo Geng, Daxin Jiang

Besides a norm-grounding knowledge model, we present a novel norm-supported ethical judgment model in line with neural module networks to alleviate dilemma situations and improve norm-level explainability.

Cultural Vocal Bursts Intensity Prediction Ethics

Dual-Personalizing Adapter for Federated Foundation Models

no code implementations28 Mar 2024 Yiyuan Yang, Guodong Long, Tao Shen, Jing Jiang, Michael Blumenstein

To address challenges in this new setting, we explore a simple yet effective solution to learn a comprehensive foundation model.

Federated Learning

Corpus-Steered Query Expansion with Large Language Models

1 code implementation28 Feb 2024 Yibin Lei, Yu Cao, Tianyi Zhou, Tao Shen, Andrew Yates

Recent studies demonstrate that query expansions generated by large language models (LLMs) can considerably enhance information retrieval systems by generating hypothetical documents that answer the queries as expansions.

Information Retrieval Retrieval

Meta-Task Prompting Elicits Embedding from Large Language Models

no code implementations28 Feb 2024 Yibin Lei, Di wu, Tianyi Zhou, Tao Shen, Yu Cao, Chongyang Tao, Andrew Yates

In this work, we introduce a new unsupervised embedding method, Meta-Task Prompting with Explicit One-Word Limitation (MetaEOL), for generating high-quality sentence embeddings from Large Language Models (LLMs) without the need for model fine-tuning or task-specific engineering.

Semantic Textual Similarity Sentence +2

A Survey on Knowledge Distillation of Large Language Models

1 code implementation20 Feb 2024 Xiaohan Xu, Ming Li, Chongyang Tao, Tao Shen, Reynold Cheng, Jinyang Li, Can Xu, DaCheng Tao, Tianyi Zhou

In the era of Large Language Models (LLMs), Knowledge Distillation (KD) emerges as a pivotal methodology for transferring advanced capabilities from leading proprietary LLMs, such as GPT-4, to their open-source counterparts like LLaMA and Mistral.

Data Augmentation Knowledge Distillation +1

Model Tailor: Mitigating Catastrophic Forgetting in Multi-modal Large Language Models

no code implementations19 Feb 2024 Didi Zhu, Zhongyi Sun, Zexi Li, Tao Shen, Ke Yan, Shouhong Ding, Kun Kuang, Chao Wu

Catastrophic forgetting emerges as a critical challenge when fine-tuning multi-modal large language models (MLLMs), where improving performance on unseen tasks often leads to a significant performance drop on the original tasks.

Image Captioning Question Answering +1

Training-time Neuron Alignment through Permutation Subspace for Improving Linear Mode Connectivity and Model Fusion

no code implementations2 Feb 2024 Zexi Li, Zhiqi Li, Jie Lin, Tao Shen, Tao Lin, Chao Wu

In deep learning, stochastic gradient descent often yields functionally similar yet widely scattered solutions in the weight space even under the same initialization, causing barriers in the Linear Mode Connectivity (LMC) landscape.

Federated Learning Linear Mode Connectivity

Leveraging Large Language Models for NLG Evaluation: A Survey

1 code implementation13 Jan 2024 Zhen Li, Xiaohan Xu, Tao Shen, Can Xu, Jia-Chen Gu, Chongyang Tao

In the rapidly evolving domain of Natural Language Generation (NLG) evaluation, introducing Large Language Models (LLMs) has opened new avenues for assessing generated content quality, e. g., coherence, creativity, and context relevance.

nlg evaluation Specificity +1

CrossBind: Collaborative Cross-Modal Identification of Protein Nucleic-Acid-Binding Residues

1 code implementation19 Dec 2023 Linglin Jing, Sheng Xu, Yifan Wang, Yuzhe Zhou, Tao Shen, Zhigang Ji, Hui Fang, Zhen Li, Siqi Sun

Accurate identification of protein nucleic-acid-binding residues poses a significant challenge with important implications for various biological processes and drug design.

Contrastive Learning Protein Language Model

Good Questions Help Zero-Shot Image Reasoning

1 code implementation4 Dec 2023 Kaiwen Yang, Tao Shen, Xinmei Tian, Xiubo Geng, Chongyang Tao, DaCheng Tao, Tianyi Zhou

QVix enables a wider exploration of visual scenes, improving the LVLMs' reasoning accuracy and depth in tasks such as visual question answering and visual entailment.

Fine-Grained Image Classification Question Answering +2

Thread of Thought Unraveling Chaotic Contexts

no code implementations15 Nov 2023 Yucheng Zhou, Xiubo Geng, Tao Shen, Chongyang Tao, Guodong Long, Jian-Guang Lou, Jianbing Shen

Large Language Models (LLMs) have ushered in a transformative era in the field of natural language processing, excelling in tasks related to text comprehension and generation.

Reading Comprehension

Enhancing Secrecy Capacity in PLS Communication with NORAN based on Pilot Information Codebooks

no code implementations2 Oct 2023 Yebo Gu, Tao Shen, Jian Song, Qingbo Wang

At the same time, selecting a NORAN design with ideal security performance from a large number of design options is also a challenging problem.

Re-Reading Improves Reasoning in Large Language Models

1 code implementation12 Sep 2023 Xiaohan Xu, Chongyang Tao, Tao Shen, Can Xu, Hongbo Xu, Guodong Long, Jian-Guang Lou

To enhance the reasoning capabilities of off-the-shelf Large Language Models (LLMs), we introduce a simple, yet general and effective prompting method, Re2, i. e., \textbf{Re}-\textbf{Re}ading the question as input.

Accurate Prediction of Antibody Function and Structure Using Bio-Inspired Antibody Language Model

1 code implementation31 Aug 2023 Hongtai Jing, Zhengtao Gao, Sheng Xu, Tao Shen, Zhangzhi Peng, Shwai He, Tao You, Shuang Ye, Wei Lin, Siqi Sun

Remarkably, BALMFold outperforms those well-established methods like AlphaFold2, IgFold, ESMFold, and OmegaFold in the antibody benchmark, demonstrating significant potential to advance innovative engineering and streamline therapeutic antibody development by reducing the need for unnecessary trials.

Language Modelling

Enhancing the Protein Tertiary Structure Prediction by Multiple Sequence Alignment Generation

2 code implementations2 Jun 2023 Le Zhang, Jiayang Chen, Tao Shen, Yu Li, Siqi Sun

The field of protein folding research has been greatly advanced by deep learning methods, with AlphaFold2 (AF2) demonstrating exceptional performance and atomic-level precision.

Language Modelling Multiple Sequence Alignment +2

Spatial-temporal Prompt Learning for Federated Weather Forecasting

no code implementations23 May 2023 Shengchao Chen, Guodong Long, Tao Shen, Tianyi Zhou, Jing Jiang

Federated weather forecasting is a promising collaborative learning framework for analyzing meteorological data across participants from different countries and regions, thus embodying a global-scale real-time weather data predictive analytics platform to tackle climate change.

Time Series Weather Forecasting

AF2-Mutation: Adversarial Sequence Mutations against AlphaFold2 on Protein Tertiary Structure Prediction

no code implementations15 May 2023 Zhongju Yuan, Tao Shen, Sheng Xu, Leiye Yu, Ruobing Ren, Siqi Sun

Deep learning-based approaches, such as AlphaFold2 (AF2), have significantly advanced protein tertiary structure prediction, achieving results comparable to real biological experimental methods.

Synergistic Interplay between Search and Large Language Models for Information Retrieval

1 code implementation12 May 2023 Jiazhan Feng, Chongyang Tao, Xiubo Geng, Tao Shen, Can Xu, Guodong Long, Dongyan Zhao, Daxin Jiang

Information retrieval (IR) plays a crucial role in locating relevant resources from vast amounts of data, and its applications have evolved from traditional knowledge bases to modern retrieval models (RMs).

Information Retrieval Retrieval

Large Language Models are Strong Zero-Shot Retriever

no code implementations27 Apr 2023 Tao Shen, Guodong Long, Xiubo Geng, Chongyang Tao, Tianyi Zhou, Daxin Jiang

In this work, we propose a simple method that applies a large language model (LLM) to large-scale retrieval in zero-shot scenarios.

Language Modelling Large Language Model +1

LexLIP: Lexicon-Bottlenecked Language-Image Pre-Training for Large-Scale Image-Text Retrieval

1 code implementation6 Feb 2023 Ziyang Luo, Pu Zhao, Can Xu, Xiubo Geng, Tao Shen, Chongyang Tao, Jing Ma, Qingwen Lin, Daxin Jiang

The conventional dense retrieval paradigm relies on encoding images and texts into dense representations using dual-stream encoders, however, it faces challenges with low retrieval speed in large-scale retrieval scenarios.

Retrieval Text Retrieval

Prompt Federated Learning for Weather Forecasting: Toward Foundation Models on Meteorological Data

1 code implementation22 Jan 2023 Shengchao Chen, Guodong Long, Tao Shen, Jing Jiang

To relieve the data exposure concern across regions, a novel federated learning approach has been proposed to collaboratively learn a brand-new spatio-temporal Transformer-based foundation model across participants with heterogeneous meteorological data.

Federated Learning Time Series +2

Iterative Proposal Refinement for Weakly-Supervised Video Grounding

no code implementations CVPR 2023 Meng Cao, Fangyun Wei, Can Xu, Xiubo Geng, Long Chen, Can Zhang, Yuexian Zou, Tao Shen, Daxin Jiang

Weakly-Supervised Video Grounding (WSVG) aims to localize events of interest in untrimmed videos with only video-level annotations.

Sentence Video Grounding

LexLIP: Lexicon-Bottlenecked Language-Image Pre-Training for Large-Scale Image-Text Sparse Retrieval

1 code implementation ICCV 2023 Ziyang Luo, Pu Zhao, Can Xu, Xiubo Geng, Tao Shen, Chongyang Tao, Jing Ma, QIngwei Lin, Daxin Jiang

To address this issue, we propose a novel sparse retrieval paradigm for ITR that exploits sparse representations in the vocabulary space for images and texts.

Image Classification Retrieval +2

Adam: Dense Retrieval Distillation with Adaptive Dark Examples

no code implementations20 Dec 2022 Chang Liu, Chongyang Tao, Xiubo Geng, Tao Shen, Dongyan Zhao, Can Xu, Binxing Jiao, Daxin Jiang

Different from previous works that only rely on one positive and hard negatives as candidate passages, we create dark examples that all have moderate relevance to the query through mixing-up and masking in discrete space.

Knowledge Distillation Retrieval

Fine-Grained Distillation for Long Document Retrieval

no code implementations20 Dec 2022 Yucheng Zhou, Tao Shen, Xiubo Geng, Chongyang Tao, Guodong Long, Can Xu, Daxin Jiang

Long document retrieval aims to fetch query-relevant documents from a large-scale collection, where knowledge distillation has become de facto to improve a retriever by mimicking a heterogeneous yet powerful cross-encoder.

Knowledge Distillation Retrieval

CCPrefix: Counterfactual Contrastive Prefix-Tuning for Many-Class Classification

no code implementations11 Nov 2022 Yang Li, Canran Xu, Guodong Long, Tao Shen, Chongyang Tao, Jing Jiang

Basically, an instance-dependent soft prefix, derived from fact-counterfactual pairs in the label space, is leveraged to complement the language verbalizers in many-class classification.

Classification counterfactual +7

DeepNoise: Signal and Noise Disentanglement based on Classifying Fluorescent Microscopy Images via Deep Learning

1 code implementation13 Sep 2022 Sen yang, Tao Shen, Yuqi Fang, Xiyue Wang, Jun Zhang, Wei Yang, Junzhou Huang, Xiao Han

The high-content image-based assay is commonly leveraged for identifying the phenotypic impact of genetic perturbations in biology field.

Disentanglement Drug Discovery +1

DUET: A Tuning-Free Device-Cloud Collaborative Parameters Generation Framework for Efficient Device Model Generalization

1 code implementation12 Sep 2022 Zheqi Lv, Wenqiao Zhang, Shengyu Zhang, Kun Kuang, Feng Wang, Yongwei Wang, Zhengyu Chen, Tao Shen, Hongxia Yang, Beng Chin Ooi, Fei Wu

DUET is deployed on a powerful cloud server that only requires the low cost of forwarding propagation and low time delay of data transmission between the device and the cloud.

Device-Cloud Collaboration Domain Adaptation +3

LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval

1 code implementation31 Aug 2022 Tao Shen, Xiubo Geng, Chongyang Tao, Can Xu, Xiaolong Huang, Binxing Jiao, Linjun Yang, Daxin Jiang

In large-scale retrieval, the lexicon-weighting paradigm, learning weighted sparse representations in vocabulary space, has shown promising results with high quality and low latency.

Language Modelling Passage Retrieval +1

LED: Lexicon-Enlightened Dense Retriever for Large-Scale Retrieval

1 code implementation29 Aug 2022 Kai Zhang, Chongyang Tao, Tao Shen, Can Xu, Xiubo Geng, Binxing Jiao, Daxin Jiang

The alignment is achieved by weakened knowledge distillations to enlighten the retriever via two aspects -- 1) a lexicon-augmented contrastive objective to challenge the dense encoder and 2) a pair-wise rank-consistent regularization to make dense model's behavior incline to the other.

Representation Learning Retrieval

Towards Robust Ranker for Text Retrieval

no code implementations16 Jun 2022 Yucheng Zhou, Tao Shen, Xiubo Geng, Chongyang Tao, Can Xu, Guodong Long, Binxing Jiao, Daxin Jiang

A ranker plays an indispensable role in the de facto 'retrieval & rerank' pipeline, but its training still lags behind -- learning from moderate negatives or/and serving as an auxiliary module for a retriever.

Passage Retrieval Retrieval +1

Interpretable RNA Foundation Model from Unannotated Data for Highly Accurate RNA Structure and Function Predictions

1 code implementation1 Apr 2022 Jiayang Chen, Zhihang Hu, Siqi Sun, Qingxiong Tan, YiXuan Wang, Qinze Yu, Licheng Zong, Liang Hong, Jin Xiao, Tao Shen, Irwin King, Yu Li

Non-coding RNA structure and function are essential to understanding various biological processes, such as cell signaling, gene expression, and post-transcriptional regulations.

Self-Supervised Learning

PCL: Peer-Contrastive Learning with Diverse Augmentations for Unsupervised Sentence Embeddings

1 code implementation28 Jan 2022 Qiyu Wu, Chongyang Tao, Tao Shen, Can Xu, Xiubo Geng, Daxin Jiang

A straightforward solution is resorting to more diverse positives from a multi-augmenting strategy, while an open question remains about how to unsupervisedly learn from the diverse positives but with uneven augmenting qualities in the text field.

Contrastive Learning Open-Ended Question Answering +3

Edge-Cloud Polarization and Collaboration: A Comprehensive Survey for AI

1 code implementation11 Nov 2021 Jiangchao Yao, Shengyu Zhang, Yang Yao, Feng Wang, Jianxin Ma, Jianwei Zhang, Yunfei Chu, Luo Ji, Kunyang Jia, Tao Shen, Anpeng Wu, Fengda Zhang, Ziqi Tan, Kun Kuang, Chao Wu, Fei Wu, Jingren Zhou, Hongxia Yang

However, edge computing, especially edge and cloud collaborative computing, are still in its infancy to announce their success due to the resource-constrained IoT scenarios with very limited algorithms deployed.

Cloud Computing Edge-computing

EventBERT: A Pre-Trained Model for Event Correlation Reasoning

no code implementations13 Oct 2021 Yucheng Zhou, Xiubo Geng, Tao Shen, Guodong Long, Daxin Jiang

Event correlation reasoning infers whether a natural language paragraph containing multiple events conforms to human common sense.

Cloze Test Common Sense Reasoning +1

Hierarchical Relation-Guided Type-Sentence Alignment for Long-Tail Relation Extraction with Distant Supervision

no code implementations Findings (NAACL) 2022 Yang Li, Guodong Long, Tao Shen, Jing Jiang

It consists of (1) a pairwise type-enriched sentence encoding module injecting both context-free and -related backgrounds to alleviate sentence-level wrong labeling, and (2) a hierarchical type-sentence alignment module enriching a sentence with the triple fact's basic attributes to support long-tail relations.

Attribute Knowledge Graphs +4

Sequential Diagnosis Prediction with Transformer and Ontological Representation

1 code implementation7 Sep 2021 Xueping Peng, Guodong Long, Tao Shen, Sen Wang, Jing Jiang

Sequential diagnosis prediction on the Electronic Health Record (EHR) has been proven crucial for predictive analytics in the medical domain.

Sequential Diagnosis

Federated Learning for Privacy-Preserving Open Innovation Future on Digital Health

no code implementations24 Aug 2021 Guodong Long, Tao Shen, Yue Tan, Leah Gerrard, Allison Clarke, Jing Jiang

Implementing an open innovation framework in the healthcare industry, namely open health, is to enhance innovation and creative capability of health-related organisations by building a next-generation collaborative framework with partner organisations and the research community.

Federated Learning Privacy Preserving

Financing Entrepreneurship and Innovation in China

no code implementations24 Aug 2021 Lin William Cong, Charles M. C. Lee, Yuanyu Qu, Tao Shen

This study reports on the current state-of-affairs in the funding of entrepreneurship and innovations in China and provides a broad survey of academic findings on the subject.

Multi-Center Federated Learning: Clients Clustering for Better Personalization

1 code implementation19 Aug 2021 Guodong Long, Ming Xie, Tao Shen, Tianyi Zhou, Xianzhi Wang, Jing Jiang, Chengqi Zhang

By comparison, a mixture of multiple global models could capture the heterogeneity across various clients if assigning the client to different global models (i. e., centers) in FL.

Clustering Decision Making +1

Dynamic Prediction Model for NOx Emission of SCR System Based on Hybrid Data-driven Algorithms

no code implementations3 Aug 2021 Zhenhao Tang, Shikui Wang, Shengxian Cao, Yang Li, Tao Shen

Aiming at the problem that delay time is difficult to determine and prediction accuracy is low in building prediction model of SCR system, a dynamic modeling scheme based on a hybrid of multiple data-driven algorithms was proposed.

feature selection FLUE +2

Improving Zero-Shot Cross-lingual Transfer for Multilingual Question Answering over Knowledge Graph

no code implementations NAACL 2021 Yucheng Zhou, Xiubo Geng, Tao Shen, Wenqiang Zhang, Daxin Jiang

That is, we can only access training data in a high-resource language, while need to answer multilingual questions without any labeled data in target languages.

Bilingual Lexicon Induction Question Answering +1

Federated Graph Learning -- A Position Paper

no code implementations24 May 2021 Huanding Zhang, Tao Shen, Fei Wu, Mingyang Yin, Hongxia Yang, Chao Wu

Federated learning (FL) is a an emerging technique that can collaboratively train a shared model while keeping the data decentralized, which is a rational solution for distributed GNN training.

Federated Learning Graph Learning +1

EBM-Fold: Fully-Differentiable Protein Folding Powered by Energy-based Models

no code implementations11 May 2021 Jiaxiang Wu, Shitong Luo, Tao Shen, Haidong Lan, Sheng Wang, Junzhou Huang

In this paper, we propose a fully-differentiable approach for protein structure optimization, guided by a data-driven generative network.

Denoising Protein Folding +1

Federated Unsupervised Representation Learning

no code implementations18 Oct 2020 Fengda Zhang, Kun Kuang, Zhaoyang You, Tao Shen, Jun Xiao, Yin Zhang, Chao Wu, Yueting Zhuang, Xiaolin Li

FURL poses two new challenges: (1) data distribution shift (Non-IID distribution) among clients would make local models focus on different categories, leading to the inconsistency of representation spaces.

Federated Learning Representation Learning

Improving Long-Tail Relation Extraction with Collaborating Relation-Augmented Attention

2 code implementations COLING 2020 Yang Li, Tao Shen, Guodong Long, Jing Jiang, Tianyi Zhou, Chengqi Zhang

Then, facilitated by the proposed base model, we introduce collaborating relation features shared among relations in the hierarchies to promote the relation-augmenting process and balance the training data for long-tail relations.

Relation Relation Extraction +1

BiteNet: Bidirectional Temporal Encoder Network to Predict Medical Outcomes

1 code implementation24 Sep 2020 Xueping Peng, Guodong Long, Tao Shen, Sen Wang, Jing Jiang, Chengqi Zhang

Electronic health records (EHRs) are longitudinal records of a patient's interactions with healthcare systems.

Clustering

Federated Mutual Learning

3 code implementations27 Jun 2020 Tao Shen, Jie Zhang, Xinkang Jia, Fengda Zhang, Gang Huang, Pan Zhou, Kun Kuang, Fei Wu, Chao Wu

The experiments show that FML can achieve better performance than alternatives in typical FL setting, and clients can be benefited from FML with different models and tasks.

Federated Learning

Self-Attention Enhanced Patient Journey Understanding in Healthcare System

1 code implementation15 Jun 2020 Xueping Peng, Guodong Long, Tao Shen, Sen Wang, Jing Jiang

The key challenge of patient journey understanding is to design an effective encoding mechanism which can properly tackle the aforementioned multi-level structured patient journey data with temporal sequential visits and a set of medical codes.

Multi-Center Federated Learning: Clients Clustering for Better Personalization

3 code implementations3 May 2020 Guodong Long, Ming Xie, Tao Shen, Tianyi Zhou, Xianzhi Wang, Jing Jiang, Chengqi Zhang

However, due to the diverse nature of user behaviors, assigning users' gradients to different global models (i. e., centers) can better capture the heterogeneity of data distributions across users.

Clustering Federated Learning

Structure-Augmented Text Representation Learning for Efficient Knowledge Graph Completion

1 code implementation30 Apr 2020 Bo Wang, Tao Shen, Guodong Long, Tianyi Zhou, Yi Chang

In experiments, we achieve state-of-the-art performance on three benchmarks and a zero-shot dataset for link prediction, with highlights of inference costs reduced by 1-2 orders of magnitude compared to a textual encoding method.

Graph Embedding Link Prediction +1

Exploiting Structured Knowledge in Text via Graph-Guided Representation Learning

no code implementations EMNLP 2020 Tao Shen, Yi Mao, Pengcheng He, Guodong Long, Adam Trischler, Weizhu Chen

In contrast to existing paradigms, our approach uses knowledge graphs implicitly, only during pre-training, to inject language models with structured knowledge via learning from raw text.

Entity Linking Knowledge Base Completion +5

Self-Attention Enhanced Selective Gate with Entity-Aware Embedding for Distantly Supervised Relation Extraction

no code implementations27 Nov 2019 Yang Li, Guodong Long, Tao Shen, Tianyi Zhou, Lina Yao, Huan Huo, Jing Jiang

Distantly supervised relation extraction intrinsically suffers from noisy labels due to the strong assumption of distant supervision.

Entity Embeddings Relation +3

Temporal Self-Attention Network for Medical Concept Embedding

1 code implementation15 Sep 2019 Xueping Peng, Guodong Long, Tao Shen, Sen Wang, Jing Jiang, Michael Blumenstein

In this paper, we propose a medical concept embedding method based on applying a self-attention mechanism to represent each medical concept.

Clustering

Tensorized Self-Attention: Efficiently Modeling Pairwise and Global Dependencies Together

2 code implementations NAACL 2019 Tao Shen, Tianyi Zhou, Guodong Long, Jing Jiang, Chengqi Zhang

Neural networks equipped with self-attention have parallelizable computation, light-weight structure, and the ability to capture both long-range and local dependencies.

Bi-Directional Block Self-Attention for Fast and Memory-Efficient Sequence Modeling

1 code implementation ICLR 2018 Tao Shen, Tianyi Zhou, Guodong Long, Jing Jiang, Chengqi Zhang

In this paper, we propose a model, called "bi-directional block self-attention network (Bi-BloSAN)", for RNN/CNN-free sequence encoding.

Reinforced Self-Attention Network: a Hybrid of Hard and Soft Attention for Sequence Modeling

1 code implementation31 Jan 2018 Tao Shen, Tianyi Zhou, Guodong Long, Jing Jiang, Sen Wang, Chengqi Zhang

In this paper, we integrate both soft and hard attention into one context fusion model, "reinforced self-attention (ReSA)", for the mutual benefit of each other.

Hard Attention Natural Language Inference +1

DiSAN: Directional Self-Attention Network for RNN/CNN-Free Language Understanding

3 code implementations14 Sep 2017 Tao Shen, Tianyi Zhou, Guodong Long, Jing Jiang, Shirui Pan, Chengqi Zhang

Recurrent neural nets (RNN) and convolutional neural nets (CNN) are widely used on NLP tasks to capture the long-term and local dependencies, respectively.

Natural Language Inference Sentence +2

Cannot find the paper you are looking for? You can Submit a new open access paper.