Topic Classification
54 papers with code • 2 benchmarks • 8 datasets
Datasets
Most implemented papers
Topic-based Evaluation for Conversational Bots
Dialog evaluation is a challenging problem, especially for non task-oriented dialogs where conversational success is not well-defined.
From Random to Supervised: A Novel Dropout Mechanism Integrated with Global Information
Dropout is used to avoid overfitting by randomly dropping units from the neural networks during training.
Topic Classification from Text Using Decision Tree, K-NN and Multinomial Naïve Bayes
One of the central motivations behind Natural Language Processing is detecting patterns.
Leap-LSTM: Enhancing Long Short-Term Memory for Text Categorization
Compared to previous models which can also skip words, our model achieves better trade-offs between performance and efficiency.
Bidirectional Context-Aware Hierarchical Attention Network for Document Understanding
The Hierarchical Attention Network (HAN) has made great strides, but it suffers a major limitation: at level 1, each sentence is encoded in complete isolation.
An Overview of the Active Gene Annotation Corpus and the BioNLP OST 2019 AGAC Track Tasks
The active gene annotation corpus (AGAC) was developed to support knowledge discovery for drug repurposing.
Sequence Labeling Approach to the Task of Sentence Boundary Detection
One of the keys to enable chatbots to communicate with human in a more natural way is the ability to handle long and complex user's utterances.
Give your Text Representation Models some Love: the Case for Basque
This is suboptimal as, for many languages, the models have been trained on smaller (or lower quality) corpora.
2kenize: Tying Subword Sequences for Chinese Script Conversion
Simplified Chinese to Traditional Chinese character conversion is a common preprocessing step in Chinese NLP.
ConCET: Entity-Aware Topic Classification for Open-Domain Conversational Agents
Our results show that ConCET significantly improves topic classification performance on both datasets, including 8-10% improvements over state-of-the-art deep learning methods.