no code implementations • EMNLP 2020 • Fei Tan, Yifan Hu, Changwei Hu, Keqian Li, Kevin Yen
In this work, we present a new language pre-training model TNT (Text Normalization based pre-training of Transformers) for content moderation.
no code implementations • EMNLP 2021 • Fei Tan, Yifan Hu, Kevin Yen, Changwei Hu
Text moderation for user generated content, which helps to promote healthy interaction among users, has been widely studied and many machine learning models have been proposed.
no code implementations • 18 Aug 2021 • Shaunak Mishra, Changwei Hu, Manisha Verma, Kevin Yen, Yifan Hu, Maxim Sviridenko
To realize this opportunity, we propose an ad text strength indicator (TSI) which: (i) predicts the click-through-rate (CTR) for an input ad text, (ii) fetches similar existing ads to create a neighborhood around the input ad, (iii) and compares the predicted CTRs in the neighborhood to declare whether the input ad is strong or weak.
no code implementations • 7 Feb 2021 • Yifan Hu, Changwei Hu, Thanh Tran, Tejaswi Kasturi, Elizabeth Joseph, Matt Gillingham
Gender information is no longer a mandatory input when registering for an account at many leading Internet companies.
no code implementations • 19 Dec 2020 • Xuan Qin, Meizhu Liu, Yifan Hu, Christina Moo, Christian M. Riblet, Changwei Hu, Kevin Yen, Haibin Ling
In this paper, we propose a method that efficiently utilizes appearance features and text vectors to accurately classify political posters from other similar political images.
no code implementations • EMNLP 2020 • Thanh Tran, Yifan Hu, Changwei Hu, Kevin Yen, Fei Tan, Kyumin Lee, Serim Park
HABERTOR inherits BERT's architecture, but is different in four aspects: (i) it generates its own vocabularies and is pre-trained from the scratch using the largest scale hatespeech dataset; (ii) it consists of Quaternion-based factorized components, resulting in a much smaller number of parameters, faster training and inferencing, as well as less memory usage; (iii) it uses our proposed multi-source ensemble heads with a pooling layer for separate input sources, to further enhance its effectiveness; and (iv) it uses a regularized adversarial training with our proposed fine-grained and adaptive noise magnitude to enhance its robustness.
no code implementations • EMNLP 2020 • Bang An, Jie Lyu, Zhenyi Wang, Chunyuan Li, Changwei Hu, Fei Tan, Ruiyi Zhang, Yifan Hu, Changyou Chen
The neural attention mechanism plays an important role in many natural language processing applications.
no code implementations • 2 Jan 2020 • Yao Zhan, Changwei Hu, Yifan Hu, Tejaswi Kasturi, Shanmugam Ramasamy, Matt Gillingham, Keith Yamamoto
In this paper, we propose graph based and deep learning models for age and gender predictions, which take into account user activities and content features.
no code implementations • 2 Jan 2020 • Changwei Hu, Yifan Hu, Sungyong Seo
The seasonality component is approximated via a non-liner function of a set of Fourier terms, and the event components are learned by a simple linear function of regressor encoding the event dates.
no code implementations • ICML 2017 • Changwei Hu, Piyush Rai, Lawrence Carin
Moreover, inference cost scales in the number of edges which is attractive for massive but sparse networks.
no code implementations • NeurIPS 2015 • Piyush Rai, Changwei Hu, Ricardo Henao, Lawrence Carin
We present a scalable Bayesian multi-label learning model based on learning low-dimensional label embeddings.
no code implementations • 18 Aug 2015 • Changwei Hu, Piyush Rai, Lawrence Carin
We present a scalable Bayesian model for low-rank factorization of massive tensors with binary observations.
no code implementations • 18 Aug 2015 • Changwei Hu, Piyush Rai, Changyou Chen, Matthew Harding, Lawrence Carin
We present a Bayesian non-negative tensor factorization model for count-valued tensor data, and develop scalable inference algorithms (both batch and online) for dealing with massive tensors.