no code implementations • ICML 2020 • Hui Ye, Zhiyu Chen, Da-Han Wang, Brian Davison
Extreme multi-label text classification (XMTC) is a task for tagging a given text with the most relevant labels from an extremely large label set.
Multi Label Text Classification Multi-Label Text Classification +1
no code implementations • 17 Oct 2024 • Zhiyu Chen, Jason Choi, Besnik Fetahu, Shervin Malmasi
We consider the task of identifying High Consideration (HC) queries.
no code implementations • 10 Sep 2024 • Zhiyu Chen, Wei Ji, Jing Xiao, Zitao Liu
Extensive experimental results on four publicly available educational datasets demonstrate the advanced predictive performance of PKT in comparison with 16 state-of-the-art models.
no code implementations • 7 Jun 2024 • Lütfi Kerem Senel, Besnik Fetahu, Davis Yoshida, Zhiyu Chen, Giuseppe Castellucci, Nikhita Vedula, Jason Choi, Shervin Malmasi
Recommender systems are widely used to suggest engaging content, and Large Language Models (LLMs) have given rise to generative recommenders.
no code implementations • 3 Apr 2024 • Parth Patwa, Simone Filice, Zhiyu Chen, Giuseppe Castellucci, Oleg Rokhlenko, Shervin Malmasi
Large Language Models (LLMs) operating in 0-shot or few-shot settings achieve competitive results in Text Classification tasks.
no code implementations • 12 Mar 2024 • Zhiyu Chen, Yu Li, Suochao Zhang, Jingbo Zhou, Jiwen Zhou, Chenfu Bao, dianhai yu
As Large Language Models (LLMs) gain great success in real-world applications, an increasing number of users are seeking to develop and deploy their customized LLMs through cloud services.
1 code implementation • 28 Jan 2024 • Dan Zhang, Yangliao Geng, Wenwen Gong, Zhongang Qi, Zhiyu Chen, Xing Tang, Ying Shan, Yuxiao Dong, Jie Tang
In this work, we investigate how to employ both batch-wise CL (BCL) and feature-wise CL (FCL) for recommendation.
no code implementations • 19 Nov 2023 • Juntao Zhang, Sheng Wang, Zhiyu Chen, Xiandi Yang, Zhiyong Peng
Finally, we develop an attention aggregator that aggregates users' preferences as the group's preferences for the group recommendation task.
no code implementations • 25 Oct 2023 • Besnik Fetahu, Zhiyu Chen, Oleg Rokhlenko, Shervin Malmasi
E-commerce product catalogs contain billions of items.
no code implementations • 20 Oct 2023 • Besnik Fetahu, Zhiyu Chen, Sudipta Kar, Oleg Rokhlenko, Shervin Malmasi
We present MULTICONER V2, a dataset for fine-grained Named Entity Recognition covering 33 entity classes across 12 languages, in both monolingual and multilingual settings.
1 code implementation • 11 Oct 2023 • Zhiyu Chen, Yujie Lu, William Yang Wang
Mental illness remains one of the most critical public health issues of our time, due to the severe scarcity and accessibility limit of professionals.
no code implementations • 6 Jun 2023 • Zhiyu Chen, Jason Choi, Besnik Fetahu, Oleg Rokhlenko, Shervin Malmasi
We propose an intent-aware FAQ retrieval system consisting of (1) an intent classifier that predicts when a user's information need can be answered by an FAQ; (2) a reformulation model that rewrites a query into a natural question.
no code implementations • 27 May 2023 • Pedro Faustini, Zhiyu Chen, Besnik Fetahu, Oleg Rokhlenko, Shervin Malmasi
Spoken Question Answering (QA) is a key feature of voice assistants, usually backed by multiple QA systems.
no code implementations • 17 May 2023 • Dan Luo, Lixin Zou, Qingyao Ai, Zhiyu Chen, Chenliang Li, Dawei Yin, Brian D. Davison
The goal of unbiased learning to rank (ULTR) is to leverage implicit user feedback for optimizing learning-to-rank systems.
no code implementations • 11 May 2023 • Besnik Fetahu, Sudipta Kar, Zhiyu Chen, Oleg Rokhlenko, Shervin Malmasi
The task highlights the need for future research on improving NER robustness on noisy data containing complex entities.
Multilingual Named Entity Recognition named-entity-recognition +3
1 code implementation • 2 May 2023 • Yujie Lu, Pan Lu, Zhiyu Chen, Wanrong Zhu, Xin Eric Wang, William Yang Wang
The key challenges of MPP are to ensure the informativeness, temporal coherence, and accuracy of plans across modalities.
no code implementations • 27 Oct 2022 • Zhiyu Chen, Jie Zhao, Anjie Fang, Besnik Fetahu, Oleg Rokhlenko, Shervin Malmasi
Furthermore, human evaluation shows that our method can generate more accurate and detailed rewrites when compared to human annotations.
1 code implementation • 18 Oct 2022 • Xinlu Zhang, Shiyang Li, Zhiyu Chen, Xifeng Yan, Linda Petzold
Our method first addresses irregularity in each single modality by (1) modeling irregular time series by dynamically incorporating hand-crafted imputation embeddings into learned interpolation embeddings via a gating mechanism, and (2) casting a series of clinical note representations as multivariate irregular time series and tackling irregularity via a time attention mechanism.
no code implementations • 13 Oct 2022 • Shiyang Li, Jianshu Chen, Yelong Shen, Zhiyu Chen, Xinlu Zhang, Zekun Li, Hong Wang, Jing Qian, Baolin Peng, Yi Mao, Wenhu Chen, Xifeng Yan
Integrating free-text explanations to in-context learning of large language models (LLM) is shown to elicit strong reasoning capabilities along with reasonable explanations.
1 code implementation • 7 Oct 2022 • Zhiyu Chen, Shiyang Li, Charese Smiley, Zhiqiang Ma, Sameena Shah, William Yang Wang
With the recent advance in large pre-trained language models, researchers have achieved record performances in NLP tasks that mostly focus on language pattern matching.
Ranked #2 on Question Answering on ConvFinQA
no code implementations • 18 Sep 2022 • Qing Jin, Zhiyu Chen, Jian Ren, Yanyu Li, Yanzhi Wang, Kaiyuan Yang
In this paper, we propose a method for training quantized networks to incorporate PIM quantization, which is ubiquitous to all PIM systems.
1 code implementation • 24 Jul 2022 • Dan Luo, Lixin Zou, Qingyao Ai, Zhiyu Chen, Dawei Yin, Brian D. Davison
Existing methods in unbiased learning to rank typically rely on click modeling or inverse propensity weighting (IPW).
no code implementations • Findings (NAACL) 2022 • Zhiyu Chen, Bing Liu, Seungwhan Moon, Chinnadhurai Sankar, Paul Crook, William Yang Wang
We also propose two new models, SimpleToDPlus and Combiner, for the proposed task.
1 code implementation • 27 Mar 2022 • Mohamed Trabelsi, Zhiyu Chen, Shuo Zhang, Brian D. Davison, Jeff Heflin
In this paper, we propose StruBERT, a structure-aware BERT model that fuses the textual and structural information of a data table to produce context-aware representations for both textual and tabular content of a data table.
1 code implementation • ICLR 2022 • Qing Jin, Jian Ren, Richard Zhuang, Sumant Hanumante, Zhengang Li, Zhiyu Chen, Yanzhi Wang, Kaiyuan Yang, Sergey Tulyakov
Our approach achieves comparable and better performance, when compared not only to existing quantization techniques with INT32 multiplication or floating-point arithmetic, but also to the full-precision counterparts, achieving state-of-the-art performance.
1 code implementation • EMNLP 2021 • Zhiyu Chen, Wenhu Chen, Charese Smiley, Sameena Shah, Iana Borova, Dylan Langdon, Reema Moussa, Matt Beane, Ting-Hao Huang, Bryan Routledge, William Yang Wang
In contrast to existing tasks on general domain, the finance domain includes complex numerical reasoning and understanding of heterogeneous representations.
Ranked #4 on Question Answering on FinQA
no code implementations • 21 Aug 2021 • Sasindu Wijeratne, Sanket Pattnaik, Zhiyu Chen, Rajgopal Kannan, Viktor Prasanna
Since developing memory controllers for different applications is time-consuming, this paper introduces a modular and programmable memory controller that can be configured for different target applications on available hardware resources.
no code implementations • 6 Jul 2021 • Zhiyu Chen, Zhanghao Yu, Qing Jin, Yan He, Jingyu Wang, Sheng Lin, Dai Li, Yanzhi Wang, Kaiyuan Yang
A compact, accurate, and bitwidth-programmable in-memory computing (IMC) static random-access memory (SRAM) macro, named CAP-RAM, is presented for energy-efficient convolutional neural network (CNN) inference.
no code implementations • 22 Jun 2021 • Youshan Zhang, Brian D. Davison, Vivien W. Talghader, Zhiyu Chen, Zhiyong Xiao, Gary J. Kunkel
To further improve segmentation results, we are the first to propose a post-processing layer to remove irrelevant portions in the segmentation result.
1 code implementation • 19 May 2021 • Yiming Sun, Feng Chen, Zhiyu Chen, Mingjie Wang
However, the perturbations of global point are not effective for misleading the victim model.
1 code implementation • 5 May 2021 • Zhiyu Chen, Shuo Zhang, Brian D. Davison
We describe the development, characteristics and availability of a test collection for the task of Web table retrieval, which uses a large-scale Web Table Corpora extracted from the Common Crawl.
1 code implementation • 12 Mar 2021 • Hanwen Zha, Zhiyu Chen, Xifeng Yan
Relation prediction in knowledge graphs is dominated by embedding based methods which mainly focus on the transductive setting.
no code implementations • 23 Feb 2021 • Mohamed Trabelsi, Zhiyu Chen, Brian D. Davison, Jeff Heflin
A variety of deep learning models have been proposed, and each model presents a set of neural network components to extract features that are used for ranking.
no code implementations • CVPR 2021 • Zhengang Li, Geng Yuan, Wei Niu, Pu Zhao, Yanyu Li, Yuxuan Cai, Xuan Shen, Zheng Zhan, Zhenglun Kong, Qing Jin, Zhiyu Chen, Sijia Liu, Kaiyuan Yang, Bin Ren, Yanzhi Wang, Xue Lin
With the increasing demand to efficiently deploy DNNs on mobile edge devices, it becomes much more important to reduce unnecessary computation and increase the execution speed.
1 code implementation • Findings (EMNLP) 2021 • Zhiyu Chen, Honglei Liu, Hu Xu, Seungwhan Moon, Hao Zhou, Bing Liu
As there is no clean mapping for a user's free form utterance to an ontology, we first model the user preferences as estimated distributions over the system ontology and map the users' utterances to such distributions.
2 code implementations • 5 Jul 2020 • Hui Ye, Zhiyu Chen, Da-Han Wang, Brian D. Davison
Extreme multi-label text classification (XMTC) is a task for tagging a given text with the most relevant labels from an extremely large label set.
1 code implementation • 19 May 2020 • Zhiyu Chen, Mohamed Trabelsi, Jeff Heflin, Yinan Xu, Brian D. Davison
Pretrained contextualized language models such as BERT have achieved impressive results on various natural language processing benchmarks.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Zhiyu Chen, Wenhu Chen, Hanwen Zha, Xiyou Zhou, Yunkai Zhang, Sairam Sundaresan, William Yang Wang
If only provided with the table, it is hard for existing models to produce controllable and high-fidelity logical generations.
1 code implementation • ACL 2020 • Wenhu Chen, Jianshu Chen, Yu Su, Zhiyu Chen, William Yang Wang
To facilitate the study of the proposed logical NLG problem, we use the existing TabFact dataset \cite{chen2019tabfact} featured with a wide range of logical/symbolic inferences as our testbed, and propose new automatic metrics to evaluate the fidelity of generation models w. r. t.\ logical inference.
2 code implementations • Findings of the Association for Computational Linguistics 2020 • Wenhu Chen, Hanwen Zha, Zhiyu Chen, Wenhan Xiong, Hong Wang, William Wang
3) a hybrid model that combines heterogeneous information to find the answer.
Ranked #4 on Question Answering on HybridQA
no code implementations • EACL 2021 • Xiyou Zhou, Zhiyu Chen, Xiaoyong Jin, William Yang Wang
We introduce HULK, a multi-task energy efficiency benchmarking platform for responsible natural language processing.
no code implementations • 3 Feb 2020 • Xinhe Jiang, Kun Wang, Kaiyi Qian, Zhaozhong Chen, Zhiyu Chen, Liangliang Lu, Lijun Xia, Fangmin Song, Shining Zhu, Xiaosong Ma
We experimentally obtain the scaling parameter of $r=-0. 88\pm$0. 03 and $-0. 78\pm$0. 07 for nonadaptive and adaptive strategies, respectively.
Quantum Physics Optics
no code implementations • 27 Jan 2020 • Zhiyu Chen, Haiyan Jia, Jeff Heflin, Brian D. Davison
We incorporate the generated schema labels into a mixed ranking model which not only considers the relevance between the query and dataset metadata but also the similarity between the query and generated schema labels.
1 code implementation • ACL 2019 • Zhiyu Chen, Hanwen Zha, Honglei Liu, Wenhu Chen, Xifeng Yan, Yu Su
Pre-trained embeddings such as word embeddings and sentence embeddings are fundamental tools facilitating a wide range of downstream NLP tasks.
Ranked #144 on Action Classification on Kinetics-400
2 code implementations • ACL 2020 • Zhiyu Chen, Harini Eavani, Wenhu Chen, Yinyin Liu, William Yang Wang
Neural-based end-to-end approaches to natural language generation (NLG) from structured data or knowledge are data-hungry, making their adoption for real-world applications difficult with limited data.
1 code implementation • NAACL 2019 • Wenhu Chen, Yu Su, Yilin Shen, Zhiyu Chen, Xifeng Yan, William Wang
Under deep neural networks, a pre-defined vocabulary is required to vectorize text inputs.
1 code implementation • 17 Sep 2018 • Chuanhai Zhang, Kurt Loken, Zhiyu Chen, Zhiyong Xiao, Gary Kunkel
Mask Editor allows drawing any bounding curve to mark objects and improves efficiency to mark objects with irregular shapes.