Task-oriented Word Embedding for Text Classification

Distributed word representation plays a pivotal role in various natural language processing tasks. In spite of its success, most existing methods only consider contextual information, which is suboptimal when used in various tasks due to a lack of task-specific features. The rational word embeddings should have the ability to capture both the semantic features and task-specific features of words. In this paper, we propose a task-oriented word embedding method and apply it to the text classification task. With the function-aware component, our method regularizes the distribution of words to enable the embedding space to have a clear classification boundary. We evaluate our method using five text classification datasets. The experiment results show that our method significantly outperforms the state-of-the-art methods.

PDF Abstract COLING 2018 PDF COLING 2018 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Text Classification AG News ToWE-SG Error 14.0 # 21
Sentiment Analysis IMDb ToWE-SG Accuracy 90.8 # 32
Sentiment Analysis SST-2 Binary classification ToWE-CBOW Accuracy 78.8 # 83

Methods


No methods listed for this paper. Add relevant methods here