no code implementations • ECCV 2020 • Tong Wang, Yousong Zhu, Chaoyang Zhao, Wei Zeng, Yao-Wei Wang, Jinqiao Wang, Ming Tang
Most of existing object detectors usually adopt a small training batch size ( ~16), which severely hinders the whole community from exploring large-scale datasets due to the extremely long training procedure.
no code implementations • ICML 2020 • Hassan Rafique, Tong Wang, Qihang Lin, Arshia Singhani
We propose a novel type of hybrid model for multi-class classification, which utilizes competing linear models to collaborate with an existing black-box model, promoting transparency in the decision-making process.
no code implementations • ACL (ECNLP) 2021 • Ying Lin, Han Wang, Jiangning Chen, Tong Wang, Yue Liu, Heng Ji, Yang Liu, Premkumar Natarajan
We first build a cross-source heterogeneous knowledge graph from customer purchase history and product knowledge graph to jointly learn customer and product embeddings.
1 code implementation • 16 Jun 2022 • Tong Wang, Guanyu Yang, Junhua Wu, Qijia He, Zhenquan Zhang
Attributed graph clustering is one of the most important tasks in graph analysis field, the goal of which is to group nodes with similar representations into the same cluster without manual guidance.
no code implementations • 28 Mar 2022 • Zimeng Li, Shichao Zhu, Bin Shao, Tie-Yan Liu, Xiangxiang Zeng, Tong Wang
Drug-drug interaction (DDI) prediction provides a drug combination strategy for systemically effective treatment.
1 code implementation • ACL 2022 • He Bai, Tong Wang, Alessandro Sordoni, Peng Shi
Class-based language models (LMs) have been long devised to address context sparsity in $n$-gram LMs.
no code implementations • 11 Mar 2022 • Ashish B. George, Tong Wang, Sergei Maslov
To understand the community structure in these energy-limited environments, we developed a microbial community consumer-resource model incorporating energetic and thermodynamic constraints on an interconnected metabolic network.
1 code implementation • 3 Feb 2022 • Jinhua Zhu, Yingce Xia, Chang Liu, Lijun Wu, Shufang Xie, Tong Wang, Yusong Wang, Wengang Zhou, Tao Qin, Houqiang Li, Tie-Yan Liu
In this work, we propose a method that directly predicts the coordinates of atoms.
no code implementations • CVPR 2022 • Tong Wang, Yousong Zhu, Yingying Chen, Chaoyang Zhao, Bin Yu, Jinqiao Wang, Ming Tang
The decision boundary between any two categories is the angular bisector of their weight vectors.
no code implementations • 13 Dec 2021 • Zhengfei Kuang, Jiaman Li, Mingming He, Tong Wang, Yajie Zhao
To make the local features aware of the global context and improve their matching accuracy, we introduce DenseGAP, a new solution for efficient Dense correspondence learning with a Graph-structured neural network conditioned on Anchor Points.
1 code implementation • 22 Nov 2021 • Tong Wang, Yuan YAO, Feng Xu, Shengwei An, Hanghang Tong, Ting Wang
We also evaluate FTROJAN against state-of-the-art defenses as well as several adaptive defenses that are designed on the frequency domain.
no code implementations • EMNLP 2021 • Trapit Bansal, Karthick Gunasekaran, Tong Wang, Tsendsuren Munkhdalai, Andrew McCallum
Meta-learning considers the problem of learning an efficient learning process that can leverage its past experience to accurately solve new tasks.
no code implementations • 14 Oct 2021 • Siyuan Liu, Yusong Wang, Tong Wang, Yifan Deng, Liang He, Bin Shao, Jian Yin, Nanning Zheng, Tie-Yan Liu
The identification of active binding drugs for target proteins (termed as drug-target interaction prediction) is the key challenge in virtual screening, which plays an essential role in drug discovery.
1 code implementation • 9 Oct 2021 • Mu Yang, Shaojin Ding, Tianlong Chen, Tong Wang, Zhangyang Wang
This work presents a lifelong learning approach to train a multilingual Text-To-Speech (TTS) system, where each language was seen as an individual task and was learned sequentially and continually.
no code implementations • 4 Jun 2021 • Chaofan Chen, Kangcheng Lin, Cynthia Rudin, Yaron Shaposhnik, Sijia Wang, Tong Wang
We propose a framework for such decisions, including a globally interpretable machine learning model, an interactive visualization of it, and several types of summaries and explanations for any given decision.
no code implementations • NAACL 2021 • Mingyue Shang, Tong Wang, Mihail Eric, Jiangning Chen, Jiyang Wang, Matthew Welch, Tiantong Deng, Akshay Grewal, Han Wang, Yue Liu, Yang Liu, Dilek Hakkani-Tur
In recent years, incorporating external knowledge for response generation in open-domain conversation systems has attracted great interest.
no code implementations • NAACL 2021 • Tong Wang, Jiangning Chen, Mohsen Malmir, Shuyan Dong, Xin He, Han Wang, Chengwei Su, Yue Liu, Yang Liu
In dialog systems, the Natural Language Understanding (NLU) component typically makes the interpretation decision (including domain, intent and slots) for an utterance before the mentioned entities are resolved.
1 code implementation • ACL 2021 • Rui Meng, Khushboo Thaker, Lei Zhang, Yue Dong, Xingdi Yuan, Tong Wang, Daqing He
Faceted summarization provides briefings of a document from different perspectives.
Ranked #1 on
Unsupervised Extractive Summarization
on FacetSum
no code implementations • 6 May 2021 • Tong Wang, Jingyi Yang, Yunyi Li, Boxiang Wang
We propose Partially Interpretable Estimators (PIE) which attribute a prediction to individual features via an interpretable model, while a (possibly) small part of the PIE prediction is attributed to the interaction of features via a black-box model, with the goal to boost the predictive performance while maintaining interpretability.
no code implementations • 6 Apr 2021 • Ying Lin, Han Wang, Jiangning Chen, Tong Wang, Yue Liu, Heng Ji, Yang Liu, Premkumar Natarajan
For example, with "add milk to my cart", a customer may refer to a certain organic product, while some customers may want to re-order products they regularly purchase.
1 code implementation • CVPR 2021 • Tong Wang, Yousong Zhu, Chaoyang Zhao, Wei Zeng, Jinqiao Wang, Ming Tang
To address the problem of long-tail distribution for the large vocabulary object detection task, existing methods usually divide the whole categories into several groups and treat each group with different strategies.
no code implementations • 4 Feb 2021 • Chit Siong Lau, Jing Yee Chee, Yee Sin Ang, Shi Wun Tong, Liemao Cao, Zi-En Ooi, Tong Wang, Lay Kee Ang, Yan Wang, Manish Chhowalla, Kuan Eng Johnson Goh
Here, temperature-dependent transfer length measurements are performed on chemical vapour deposition grown single-layer and bilayer WS$_2$ devices with indium alloy contacts.
Materials Science Mesoscale and Nanoscale Physics
no code implementations • 28 Jan 2021 • Alexei V. Tkachenko, Sergei Maslov, Tong Wang, Ahmed Elbanna, George N. Wong, Nigel Goldenfeld
It is well recognized that population heterogeneity plays an important role in the spread of epidemics.
no code implementations • 12 Jan 2021 • Shaosheng Xu, Jinde Cao, Yichao Cao, Tong Wang
As gradient descent method in deep learning causes a series of questions, this paper proposes a novel gradient-free deep learning structure.
no code implementations • 6 Jan 2021 • Yao Li, Tong Wang, Juanrong Zhang, Bin Shao, Haipeng Gong, Yusong Wang, Siyuan Liu, Tie-Yan Liu
We performed molecular dynamics simulation on the S protein with a focus on the function of its N-terminal domains (NTDs).
no code implementations • 17 Nov 2020 • Tong Wang, Maytal Saar-Tsechansky
We formulate a multi-objective optimization for building a surrogate model, where we simultaneously optimize for both predictive performance and bias.
no code implementations • NAACL 2021 • Rui Meng, Xingdi Yuan, Tong Wang, Sanqiang Zhao, Adam Trischler, Daqing He
Recent years have seen a flourishing of neural keyphrase generation (KPG) works, including the release of several large-scale datasets and a host of new models to tackle them.
no code implementations • 19 Jul 2020 • Xinwei Chen, Tong Wang, Barrett W. Thomas, Marlin W. Ulmer
The demand for same-day delivery (SDD) has increased rapidly in the last few years and has particularly boomed during the COVID-19 pandemic.
1 code implementation • 3 Jul 2020 • Dat Hong, Stephen S. Baek, Tong Wang
We propose a novel interpretable deep neural network for text classification, called ProtoryNet, based on a new concept of prototype trajectories.
1 code implementation • EMNLP 2020 • Tu Vu, Tong Wang, Tsendsuren Munkhdalai, Alessandro Sordoni, Adam Trischler, Andrew Mattarella-Micke, Subhransu Maji, Mohit Iyyer
We also develop task embeddings that can be used to predict the most transferable source tasks for a given target task, and we validate their effectiveness in experiments controlled for source and target data size.
no code implementations • 1 Apr 2020 • Fengling Li, Tong Wang, Lei Zhu, Zheng Zhang, Xinhua Wang
Unlike previous cross-modal hashing approaches, our learning framework jointly optimizes semantic preserving that transforms deep features of multimedia data into binary hash codes, and the semantic regression which directly regresses query modality representation to explicit label.
no code implementations • 23 Feb 2020 • Azhar Hussain, Tong Wang, Cao Jiahua
We consider a system to optimize duration of traffic signals using multi-agent deep reinforcement learning and Vehicle-to-Everything (V2X) communication.
no code implementations • 19 Feb 2020 • Chen Liao, Tong Wang, Sergei Maslov, Joao B. Xavier
We used the three models to study each community's limits of robustness to perturbations such as variations in resource supply, antibiotic treatments and invasion by other "cheaters" species.
no code implementations • 10 Feb 2020 • Danqing Pan, Tong Wang, Satoshi Hara
We present an interpretable companion model for any pre-trained black-box classifiers.
no code implementations • 9 Nov 2019 • Tong Wang, Fujie Jin, Yu, Hu, Yuan Cheng
The prediction model and the interpretable insights can be applied to assist fundraisers with better promoting their fundraising campaigns and can potentially help crowdfunding platforms to provide more timely feedback to all fundraisers.
no code implementations • 23 Sep 2019 • Hassan Rafique, Tong Wang, Qihang Lin
Driven by an increasing need for model interpretability, interpretable models have become strong competitors for black-box models in many real applications.
1 code implementation • 9 Sep 2019 • Rui Meng, Xingdi Yuan, Tong Wang, Peter Brusilovsky, Adam Trischler, Daqing He
Recently, concatenating multiple keyphrases as a target sequence has been proposed as a new learning paradigm for keyphrase generation.
1 code implementation • NeurIPS 2019 • Tsendsuren Munkhdalai, Alessandro Sordoni, Tong Wang, Adam Trischler
We augment recurrent neural networks with an external memory mechanism that builds upon recent progress in metalearning.
no code implementations • 10 May 2019 • Tong Wang, Qihang Lin
The interpretable model substitutes the black-box model on a subset of data where the black-box is overkill or nearly overkill, gaining transparency at no or low cost of the predictive accuracy.
no code implementations • 13 Dec 2018 • Yunyi Li, Tong Wang
Our goal is to predict the location of the next crime in a crime series, based on the identified previous offenses in the series.
1 code implementation • NeurIPS 2018 • Tong Wang
We present the Multi-value Rule Set (MRS) for interpretable classification with feature efficient presentations.
no code implementations • 30 Nov 2018 • Chaofan Chen, Kangcheng Lin, Cynthia Rudin, Yaron Shaposhnik, Sijia Wang, Tong Wang
We propose a possible solution to a public challenge posed by the Fair Isaac Corporation (FICO), which is to provide an explainable model for credit risk assessment.
1 code implementation • ACL 2020 • Xingdi Yuan, Tong Wang, Rui Meng, Khushboo Thaker, Peter Brusilovsky, Daqing He, Adam Trischler
With both previous and new evaluation metrics, our model outperforms strong baselines on all datasets.
1 code implementation • 6 Jul 2018 • Tong Wang, Veerajalandhar Allareddy, Sankeerth Rampa, Veerasathpurush Allareddy
We propose a Bayesian framework for formulating a MRS model and propose an efficient inference method for learning a maximum \emph{a posteriori}, incorporating theoretically grounded bounds to iteratively reduce the search space and improve the search efficiency.
no code implementations • WS 2018 • S Subramanian, eep, Tong Wang, Xingdi Yuan, Saizheng Zhang, Adam Trischler, Yoshua Bengio
We propose a two-stage neural model to tackle question generation from documents.
1 code implementation • 12 Feb 2018 • Tong Wang
This work addresses the situation where a black-box model with good predictive performance is chosen over its interpretable competitors, and we show interpretability is still achievable in this case.
no code implementations • 16 Oct 2017 • Tong Wang, Cynthia Rudin
The Bayesian model has tunable parameters that can characterize subgroups with various sizes, providing users with more flexible choices of models from the \emph{treatment efficient frontier}.
no code implementations • 15 Oct 2017 • Tong Wang
MARS introduces a more generalized form of association rules that allows multiple values in a condition.
no code implementations • LREC 2018 • Boyang Li, Beth Cardier, Tong Wang, Florian Metze
Stories are a vital form of communication in human culture; they are employed daily to persuade, to elicit sympathy, or to convey a message.
no code implementations • 8 Jul 2017 • Tong Wang, Ping Chen, Boyang Li
An important and difficult challenge in building computational models for narratives is the automatic evaluation of narrative quality.
no code implementations • 14 Jun 2017 • Sandeep Subramanian, Tong Wang, Xingdi Yuan, Saizheng Zhang, Yoshua Bengio, Adam Trischler
We propose a two-stage neural model to tackle question generation from documents.
no code implementations • 5 Jun 2017 • Tong Wang, Xingdi Yuan, Adam Trischler
We propose a generative machine comprehension model that learns jointly to ask and answer questions based on documents.
4 code implementations • WS 2017 • Xingdi Yuan, Tong Wang, Caglar Gulcehre, Alessandro Sordoni, Philip Bachman, Sandeep Subramanian, Saizheng Zhang, Adam Trischler
We propose a recurrent neural model that generates natural-language questions from documents, conditioned on answers.
no code implementations • 21 Apr 2017 • Ping Chen, Fei Wu, Tong Wang, Wei Ding
In this paper, we will present some preliminary results on one especially useful and challenging problem in NLP system evaluation: how to pinpoint content differences of two text passages (especially for large pas-sages such as articles and books).
1 code implementation • WS 2017 • Adam Trischler, Tong Wang, Xingdi Yuan, Justin Harris, Alessandro Sordoni, Philip Bachman, Kaheer Suleman
We present NewsQA, a challenging machine comprehension dataset of over 100, 000 human-generated question-answer pairs.
11 code implementations • 28 Nov 2016 • Payal Bajaj, Daniel Campos, Nick Craswell, Li Deng, Jianfeng Gao, Xiaodong Liu, Rangan Majumder, Andrew McNamara, Bhaskar Mitra, Tri Nguyen, Mir Rosenberg, Xia Song, Alina Stoica, Saurabh Tiwary, Tong Wang
The size of the dataset and the fact that the questions are derived from real user search queries distinguishes MS MARCO from other well-known publicly available datasets for machine reading comprehension and question-answering.
1 code implementation • 27 Sep 2016 • Jipeng Qiang, Ping Chen, Tong Wang, Xindong Wu
Inferring topics from the overwhelming amount of short texts becomes a critical but challenging task for many content analysis tasks, such as content charactering, user interest profiling, and emerging topic detecting.
no code implementations • 13 Sep 2016 • Tong Wang, Ping Chen, Kevin Amaral, Jipeng Qiang
Text simplification (TS) aims to reduce the lexical and structural complexity of a text, while still retaining the semantic meaning.
no code implementations • 6 Nov 2015 • Tong Wang, Cynthia Rudin
Or's of And's (OA) models are comprised of a small number of disjunctions of conjunctions, also called disjunctive normal form.
no code implementations • 28 Apr 2015 • Tong Wang, Cynthia Rudin, Finale Doshi-Velez, Yimin Liu, Erica Klampfl, Perry MacNeille
In both cases, there are prior parameters that the user can set to encourage the model to have a desired size and shape, to conform with a domain-specific definition of interpretability.