1 code implementation • 20 Sep 2023 • Wenhang Shi, Yiren Chen, Zhe Zhao, Wei Lu, Kimmo Yan, Xiaoyong Du
Therefore, we shift the attention to the current task learning stage, presenting a novel framework, C&F (Create and Find Flatness), which builds a flat training space for each task in advance.
no code implementations • 24 Aug 2023 • Guangyu Chen, Yu Wu, Shujie Liu, Tao Liu, Xiaoyong Du, Furu Wei
Recent breakthroughs in zero-shot voice synthesis have enabled imitating a speaker's voice using just a few seconds of recording while maintaining a high level of realism.
1 code implementation • 3 Jul 2023 • Cheng Chen, Yong Wang, Lizi Liao, Yueguo Chen, Xiaoyong Du
Given a limited labeling budget, active learning (AL) aims to sample the most informative instances from an unlabeled pool to acquire labels for subsequent model training.
1 code implementation • 15 Jun 2023 • Zihui Gu, Ju Fan, Nan Tang, Songyue Zhang, Yuxin Zhang, Zui Chen, Lei Cao, Guoliang Li, Sam Madden, Xiaoyong Du
PLMs can perform well in schema alignment but struggle to achieve complex reasoning, while LLMs is superior in complex reasoning tasks but cannot achieve precise schema alignment.
1 code implementation • SIGMOD/PODS 2023 • Jianhong Tu, Ju Fan, Nan Tang, Peng Wang, Guoliang Li, Xiaoyong Du, Xiaofeng Jia, Song Gao
The widely used practice is to build task-specific or even dataset-specific solutions, which are hard to generalize and disable the opportunities of knowledge sharing that can be learned from different datasets and multiple tasks.
no code implementations • 7 Apr 2023 • Sibei Chen, Hanbing Liu, Weiting Jin, Xiangyu Sun, Xiaoyao Feng, Ju Fan, Xiaoyong Du, Nan Tang
Orchestrating a high-quality data preparation program is essential for successful machine learning (ML), but it is known to be time and effort consuming.
3 code implementations • 13 Dec 2022 • Zhe Zhao, Yudong Li, Cheng Hou, Jing Zhao, Rong Tian, Weijie Liu, Yiren Chen, Ningyuan Sun, Haoyan Liu, Weiquan Mao, Han Guo, Weigang Guo, Taiqiang Wu, Tao Zhu, Wenhang Shi, Chen Chen, Shan Huang, Sihong Chen, Liqun Liu, Feifei Li, Xiaoshuai Chen, Xingwu Sun, Zhanhui Kang, Xiaoyong Du, Linlin Shen, Kimmo Yan
The proposed pre-training models of different modalities are showing a rising trend of homogeneity in their model structures, which brings the opportunity to implement different pre-training models within a uniform framework.
1 code implementation • 5 Nov 2022 • Zihui Gu, Ju Fan, Nan Tang, Preslav Nakov, Xiaoman Zhao, Xiaoyong Du
In particular, on the complex set of TabFact, which contains multiple operations, PASTA largely outperforms the previous state of the art by 4. 7 points (85. 6% vs. 80. 9%), and the gap between PASTA and human performance on the small TabFact test set is narrowed to just 1. 5 points (90. 6% vs. 92. 1%).
Ranked #2 on
Table-based Fact Verification
on TabFact
no code implementations • 17 Sep 2022 • Zhaoxin Fan, Fengxin Li, Hongyan Liu, Jun He, Xiaoyong Du
In this paper, we research the new topic of object effects recommendation in micro-video platforms, which is a challenging but important task for many practical applications such as advertisement insertion.
1 code implementation • SIGMOD/PODS 2022 • Jianhong Tu, Ju Fan, Nan Tang, Peng Wang, Chengliang Chai, Guoliang Li, Ruixue Fan, Xiaoyong Du
Entity resolution (ER) is a core problem of data integration.
Ranked #2 on
Entity Resolution
on WDC Watches-small
no code implementations • 29 Aug 2021 • Zhaoxin Fan, Zhenbo Song, Wenping Zhang, Hongyan Liu, Jun He, Xiaoyong Du
Third, we apply these kernels to previous point cloud features to generate new features, which is the well-known SO(3) mapping process.
no code implementations • 1 May 2021 • Zhaoxin Fan, Zhenbo Song, Hongyan Liu, Zhiwu Lu, Jun He, Xiaoyong Du
Point cloud-based large scale place recognition is fundamental for many applications like Simultaneous Localization and Mapping (SLAM).
no code implementations • 1 Jan 2021 • Chuyuan Xiong, Deyuan Zhang, Tao Liu, Xiaoyong Du, Jiankun Tian, Songyan Xue
In this paper, a baseline evaluation framework is proposed for voice-face matching and retrieval tasks.
no code implementations • 4 Dec 2020 • Nan Tang, Ju Fan, Fangyi Li, Jianhong Tu, Xiaoyong Du, Guoliang Li, Sam Madden, Mourad Ouzzani
RPT is pre-trained for a tuple-to-tuple model by corrupting the input tuple and then learning a model to reconstruct the original tuple.
1 code implementation • NeurIPS 2020 • Ming Chen, Zhewei Wei, Bolin Ding, Yaliang Li, Ye Yuan, Xiaoyong Du, Ji-Rong Wen
Most notably, GBP can deliver superior performance on a graph with over 60 million nodes and 1. 8 billion edges in less than half an hour on a single machine.
1 code implementation • 28 Aug 2020 • Ju Fan, Tongyu Liu, Guoliang Li, Junyou Chen, Yuwei Shen, Xiaoyong Du
We conduct extensive experiments to explore the design space and compare with traditional data synthesis approaches.
1 code implementation • 24 Dec 2019 • Zhiqiang Gong, Weidong Hu, Xiaoyong Du, Ping Zhong, Panhe Hu
Deep learning methods have played a more and more important role in hyperspectral image classification.
no code implementations • 21 Nov 2019 • Chuyuan Xiong, Deyuan Zhang, Tao Liu, Xiaoyong Du
It achieves state-of-the-art performance with various performance metrics on different tasks and with high test confidence on large scale datasets, which can be taken as a baseline for the follow-up research.
1 code implementation • IJCNLP 2019 • Zhe Zhao, Hui Chen, Jinbin Zhang, Xin Zhao, Tao Liu, Wei Lu, Xi Chen, Haotang Deng, Qi Ju, Xiaoyong Du
Existing works, including ELMO and BERT, have revealed the importance of pre-training for NLP tasks.
no code implementations • WS 2018 • Bofang Li, Aleks Drozd, R, Tao Liu, Xiaoyong Du
Subword-level information is crucial for capturing the meaning and morphology of words, especially for out-of-vocabulary entries.
2 code implementations • ACL 2018 • Shen Li, Zhe Zhao, Renfen Hu, Wensi Li, Tao Liu, Xiaoyong Du
Analogical reasoning is effective in capturing linguistic regularities.
no code implementations • EMNLP 2017 • Shen Li, Zhe Zhao, Tao Liu, Renfen Hu, Xiaoyong Du
Convolutional Neural Networks (CNNs) are widely used in NLP tasks.
no code implementations • EMNLP 2017 • Bofang Li, Tao Liu, Zhe Zhao, Buzhou Tang, Aleks Drozd, R, Anna Rogers, Xiaoyong Du
The number of word embedding models is growing every year.
no code implementations • EMNLP 2017 • Zhe Zhao, Tao Liu, Shen Li, Bofang Li, Xiaoyong Du
The existing word representation methods mostly limit their information source to word co-occurrence statistics.
1 code implementation • COLING 2016 • Bofang Li, Zhe Zhao, Tao Liu, Puwei Wang, Xiaoyong Du
We train n-gram embeddings and use NB weighting to guide the neural models to focus on important words.
1 code implementation • 27 Dec 2015 • Bofang Li, Tao Liu, Xiaoyong Du, Deyuan Zhang, Zhe Zhao
Many document embeddings methods have been proposed to capture semantics, but they still can't outperform bag-of-ngram based methods on this task.