1 code implementation • EMNLP 2020 • Xiangji Zeng, Yunliang Li, Yuchen Zhai, Yin Zhang
In this paper, we decompose the sentence into two parts: entity and context, and rethink the relationship between them and model performance from a causal perspective.
no code implementations • EMNLP 2021 • Chengyu Wang, Haojie Pan, Minghui Qiu, Jun Huang, Fei Yang, Yin Zhang
For tasks related to distant domains with different class label sets, PLMs may memorize non-transferable knowledge for the target domain and suffer from negative transfer.
no code implementations • EMNLP 2021 • Haiwen Hong, Jingfeng Zhang, Yin Zhang, Yao Wan, Yulei Sui
Obviously, unchanged fix is not the correct fix because it is the same as the buggy code that needs to be fixed.
no code implementations • COLING 2022 • Guodun Li, Yuchen Zhai, Qianglong Chen, Xing Gao, Ji Zhang, Yin Zhang
Intent detection is at the core of task-oriented dialogue systems.
no code implementations • Findings (EMNLP) 2021 • Zhi Li, Yuchen Zhai, Chengyu Wang, Minghui Qiu, Kailiang Li, Yin Zhang
Inspired by the fact that words with similar semantic can share a part of weights, we divide the embeddings of words into two parts: unique embedding and class embedding.
1 code implementation • 23 Dec 2024 • Yue Deng, Yan Yu, Weiyu Ma, ZiRui Wang, Wenhui Zhu, Jian Zhao, Yin Zhang
SMAC-HARD supports customizable opponent strategies, randomization of adversarial policies, and interfaces for MARL self-play, enabling agents to generalize to varying opponent behaviors and improve model stability.
no code implementations • 15 Nov 2024 • Xingxi Yin, Zhi Li, Jingfeng Zhang, Chenglin Li, Yin Zhang
Text-to-image (T2I) diffusion models, with their impressive generative capabilities, have been adopted for image editing tasks, demonstrating remarkable efficacy.
no code implementations • 14 Nov 2024 • Chenglin Li, Qianglong Chen, Zhi Li, Feng Tao, Yin Zhang
Recent advancements in Large Video-Language Models (LVLMs) have driven the development of benchmarks designed to assess cognitive abilities in video-based tasks.
1 code implementation • 13 Nov 2024 • Peng Wang, Lingzhe Zhao, Yin Zhang, Shiyu Zhao, Peidong Liu
In our experiments, we demonstrate that MBA-SLAM surpasses previous state-of-the-art methods in both camera localization and map reconstruction, showcasing superior performance across a range of datasets, including synthetic and real datasets featuring sharp images as well as those affected by motion blur, highlighting the versatility and robustness of our approach.
1 code implementation • 21 Oct 2024 • Yue Deng, Weiyu Ma, Yuxin Fan, Yin Zhang, Haifeng Zhang, Jian Zhao
StarCraft Multi-Agent Challenge (SMAC) is one of the most commonly used experimental environments in multi-agent reinforcement learning (MARL), where the specific task is to control a set number of allied units to defeat enemy forces.
no code implementations • 14 Oct 2024 • Chenglin Li, Qianglong Chen, Zhi Li, Feng Tao, Yicheng Li, Hao Chen, Fei Yu, Yin Zhang
With tree search and evaluation models, it can efficiently guide each instruction to evolve into a high-quality form, aiding in instruction fine-tuning.
1 code implementation • 7 Aug 2024 • Shihao Tu, Yupeng Zhang, Jing Zhang, Zhendong Fu, Yin Zhang, Yang Yang
PowerPM consists of a temporal encoder and a hierarchical encoder.
no code implementations • 31 Jul 2024 • Taiji Li, Zhi Li, Yin Zhang
Specifically, SliSum divides the source article into overlapping windows, and utilizes LLM to generate local summaries for the content in the windows.
no code implementations • 20 Jun 2024 • Junjie Wang, Yin Zhang, Yatai Ji, Yuxiang Zhang, Chunyang Jiang, YuBo Wang, Kang Zhu, Zekun Wang, Tiezhen Wang, Wenhao Huang, Jie Fu, Bei Chen, Qunshu Lin, Minghao Liu, Ge Zhang, Wenhu Chen
Recent advancements in Large Multimodal Models (LMMs) have leveraged extensive multimodal datasets to enhance capabilities in complex knowledge-driven tasks.
no code implementations • 16 Apr 2024 • Ruifeng Li, Dongzhan Zhou, Ancheng Shen, Ao Zhang, Mao Su, Mingqian Li, Hongyang Chen, Gang Chen, Yin Zhang, Shufei Zhang, Yuqiang Li, Wanli Ouyang
Overall, our work illustrates the benefits and potential of using PEMAL in AIDD and other scenarios with data scarcity and noise.
1 code implementation • 25 Mar 2024 • Yin Zhang, Jinhong Deng, Peidong Liu, Wen Li, Shiyu Zhao
A new benchmark for cross-domain MAV detection is proposed based on the proposed dataset.
no code implementations • 29 Feb 2024 • Jianyu Guan, Zongming Yin, Tianyi Zhang, Leihui Chen, Yin Zhang, Fei Huang, Jufeng Chen, Shuguang Han
In the end, the extracted common knowledge is adopted for target entity model training.
no code implementations • 8 Jan 2024 • Zhongjiang He, Zihan Wang, Xinzhang Liu, Shixuan Liu, Yitong Yao, Yuyao Huang, Xuelong Li, Yongxiang Li, Zhonghao Che, Zhaoxi Zhang, Yan Wang, Xin Wang, Luwen Pu, Huinan Xu, Ruiyu Fang, Yu Zhao, Jie Zhang, Xiaomeng Huang, Zhilong Lu, Jiaxin Peng, Wenjun Zheng, Shiquan Wang, Bingkai Yang, Xuewei he, Zhuoru Jiang, Qiyi Xie, Yanhan Zhang, Zhongqiu Li, Lingling Shi, Weiwei Fu, Yin Zhang, Zilu Huang, Sishi Xiong, Yuxiang Zhang, Chao Wang, Shuangyong Song
Subsequently, the model undergoes fine-tuning to align with human preferences, following a detailed methodology that we describe.
2 code implementations • 18 Dec 2023 • Hanqing Guo, Ye Zheng, Yin Zhang, Zhi Gao, Shiyu Zhao
In this paper, we propose a global-local MAV detector that can fuse both motion and appearance features for MAV detection under challenging conditions.
no code implementations • 17 Dec 2023 • Chenglin Li, Qianglong Chen, Liangyue Li, Caiyu Wang, Yicheng Li, Zulong Chen, Yin Zhang
While large language models (LLMs) have demonstrated exceptional performance in recent natural language processing (NLP) tasks, their deployment poses substantial challenges due to high computational and memory demands in real-world applications.
no code implementations • 11 Dec 2023 • Yueyao Yu, Yin Zhang
Since its introduction in 2017, Transformer has emerged as the leading neural network architecture, catalyzing revolutionary advancements in many AI disciplines.
1 code implementation • 13 Oct 2023 • Junwen Zhang, Yin Zhang
To alleviate this issue, we present a novel Hierarchical Contrastive Learning Framework (HiCL) for zero-shot slot filling.
no code implementations • 22 Sep 2023 • Chenin Li, Qianglong Chen, Yin Zhang, Yifei Zhang, Hongxiang Yao
Commonsense reasoning is a pivotal skill for large language models, yet it presents persistent challenges in specific tasks requiring this competence.
no code implementations • 20 Aug 2023 • Chenwei Wang, Siyi Luo, Yulin Huang, Jifang Pei, Yin Zhang, Jianyu Yang
The designed augmenter increases the amount of information available for supervised training and improves the separability of the extracted features.
no code implementations • 20 Aug 2023 • Chenwei Wang, Jifang Pei, Siyi Luo, Weibo Huo, Yulin Huang, Yin Zhang, Jianyu Yang
Therefore, we proposed a SAR ship recognition method via multi-scale feature attention and adaptive-weighted classifier to enhance features in each scale, and adaptively choose the effective feature scale for accurate recognition.
no code implementations • 20 Aug 2023 • Chenwei Wang, Siyi Luo, Jifang Pei, Yulin Huang, Yin Zhang, Jianyu Yang
However, the characteristics of SAR ship images, large inner-class variance, and small interclass difference lead to the whole features containing useless partial features and a single feature center for each class in the classifier failing with large inner-class variance.
no code implementations • 20 Aug 2023 • Chenwei Wang, Siyi Luo, Jifang Pei, Xiaoyu Liu, Yulin Huang, Yin Zhang, Jianyu Yang
In this letter, we propose an entropy-awareness meta-learning method that improves the exclusiveness of feature distribution of known classes which means our method is effective for not only classifying the seen classes but also encountering the unseen other classes.
1 code implementation • 20 Aug 2023 • Chenwei Wang, Siyi Luo, Jifang Pei, Yulin Huang, Yin Zhang, Jianyu Yang
Based on the initial recognition results, the feature capture module automatically searches and locks the crucial image regions for correct recognition, which we named as the golden key of image.
1 code implementation • 18 Aug 2023 • Chenwei Wang, You Qin, Li Li, Siyi Luo, Yulin Huang, Jifang Pei, Yin Zhang, Jianyu Yang
As a result, it has a detrimental causal effect damaging the efficacy of feature $X$ extracted from SAR images, leading to weak generalization of SAR ATR with limited data.
no code implementations • 10 Aug 2023 • Chenwei Wang, Yulin Huang, Xiaoyu Liu, Jifang Pei, Yin Zhang, Jianyu Yang
Convolutional neural networks (CNNs) have dominated the synthetic aperture radar (SAR) automatic target recognition (ATR) for years.
no code implementations • 10 Aug 2023 • Chenwei Wang, Jifang Pei, Xiaoyu Liu, Yulin Huang, Deqing Mao, Yin Zhang, Jianyu Yang
The similarity discriminator can differentiate the generated SAR target images from the real SAR images to ensure the accuracy of the generated, while the azimuth predictor measures the difference of azimuth between the generated and the desired to ensure the azimuth controllability of the generated.
1 code implementation • 27 Jun 2023 • Chenwei Wang, Siyi Luo, Lin Liu, Yin Zhang, Jifang Pei, Yulin Huang, Jianyu Yang
In recent years, deep learning has been widely used to solve the bottleneck problem of synthetic aperture radar (SAR) automatic target recognition (ATR).
3 code implementations • 1 Jun 2023 • Wenjin Wang, Yunhao Li, Yixin Ou, Yin Zhang
Instead, in this paper, we find that instruction-tuning language models like Claude and ChatGPT can understand layout by spaces and line breaks.
Ranked #9 on Visual Question Answering (VQA) on DocVQA test
1 code implementation • 23 May 2023 • Bo Zhou, Qianglong Chen, Tianyu Wang, Xiaomi Zhong, Yin Zhang
To fully evaluate the overall performance of different NLP models in a given domain, many evaluation benchmarks are proposed, such as GLUE, SuperGLUE and CLUE.
no code implementations • 14 May 2023 • Qianglong Chen, Guohai Xu, Ming Yan, Ji Zhang, Fei Huang, Luo Si, Yin Zhang
Existing knowledge-enhanced methods have achieved remarkable results in certain QA tasks via obtaining diverse knowledge from different knowledge bases.
no code implementations • 13 May 2023 • Qianglong Chen, Feng Ji, Feng-Lin Li, Guohai Xu, Ming Yan, Ji Zhang, Yin Zhang
To support cost-effective language inference in multilingual settings, we propose AMTSS, an adaptive multi-teacher single-student distillation framework, which allows distilling knowledge from multiple teachers to a single student.
1 code implementation • CVPR 2023 • Wenjin Wang, Yunqing Hu, Qianglong Chen, Yin Zhang
In this paper, we propose the Parameter Allocation & Regularization (PAR), which adaptively select an appropriate strategy for each task from parameter allocation and regularization based on its learning difficulty.
1 code implementation • 26 Oct 2022 • Jie Cao, Yin Zhang
However, such models can't predict a relatively complete and variable-length label subset for each document, because they select positive labels relevant to the document by a fixed threshold or take top k labels in descending order of scores.
Multi Label Text Classification Multi-Label Text Classification +1
no code implementations • 25 Oct 2022 • Yin Zhang, Ruoxi Wang, Tiansheng Yao, Xinyang Yi, Lichan Hong, James Caverlee, Ed H. Chi, Derek Zhiyuan Cheng
In this work, we aim to improve tail item recommendations while maintaining the overall performance with less training and serving cost.
2 code implementations • 12 Oct 2022 • Qiming Peng, Yinxu Pan, Wenjin Wang, Bin Luo, Zhenyu Zhang, Zhengjie Huang, Teng Hu, Weichong Yin, Yongfeng Chen, Yin Zhang, Shikun Feng, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang
Recent years have witnessed the rise and success of pre-training techniques in visually-rich document understanding.
Ranked #2 on Semantic entity labeling on FUNSD
1 code implementation • 27 Sep 2022 • Shen Huang, Yuchen Zhai, Xinwei Long, Yong Jiang, Xiaobin Wang, Yin Zhang, Pengjun Xie
Speech Entity Linking aims to recognize and disambiguate named entities in spoken languages.
no code implementations • 18 Sep 2022 • Wenjin Wang, Zhengjie Huang, Bin Luo, Qianglong Chen, Qiming Peng, Yinxu Pan, Weichong Yin, Shikun Feng, Yu Sun, dianhai yu, Yin Zhang
At first, a document graph is proposed to model complex relationships among multi-grained multimodal elements, in which salient visual regions are detected by a cluster-based method.
no code implementations • 6 Aug 2022 • Yin Zhang, Can Xu, XianJun Wu, Yan Zhang, LiGang Dong, Weigang Wang
Recently, many efforts have been devoted to improving Tag-aware recommendation systems (TRS) with Graph Convolutional Networks (GCN), which has become new state-of-the-art for the general recommendation.
no code implementations • 1 Aug 2022 • Qianglong Chen, Feng-Lin Li, Guohai Xu, Ming Yan, Ji Zhang, Yin Zhang
We evaluate our approach on a variety of knowledge driven and language understanding tasks, including NER, relation extraction, CommonsenseQA, OpenBookQA and GLUE.
1 code implementation • 6 Jul 2022 • Qianglong Chen, Xiangji Zeng, Jiangang Zhu, Yin Zhang, Bojia Lin, Yang Yang, Daxin Jiang
Gazetteer is widely used in Chinese named entity recognition (NER) to enhance span boundary detection and type classification.
no code implementations • 17 May 2022 • Xinyu Chen, Renjie Li, Yueyao Yu, Yuanwen Shen, Wenye Li, Zhaoyu Zhang, Yin Zhang
In this work, we propose the first-ever Transformer model (POViT) to efficiently design and simulate semiconductor photonic devices with multiple objectives.
no code implementations • 22 Apr 2022 • Yunqing Hu, Xuan Jin, Yin Zhang, Haiwen Hong, Jingfeng Zhang, Feihu Yan, Yuan He, Hui Xue
Finally, we propose a weakly supervised object localization-based approach to extract multi-scale local features, to form a multi-view pipeline.
Multi-Label Image Recognition Weakly-Supervised Object Localization
no code implementations • 29 Sep 2021 • Yueyao Yu, Yin Zhang
We introduce a notion of variability to view such issues under the setting of a fixed number of parameters which is, in general, a dominant cost-factor.
no code implementations • 26 Aug 2021 • Guodun Li, Yuchen Zhai, Zehao Lin, Yin Zhang
Second, we construct the plugable multi-modal scene retriever to retrieve scenes represented with pairs of an image and its stylized caption, which are similar to the query image or caption in the large-scale factual data.
no code implementations • ACL 2021 • Qianglong Chen, Feng Ji, Xiangji Zeng, Feng-Lin Li, Ji Zhang, Haiqing Chen, Yin Zhang
In order to better understand the reason behind model behaviors (i. e., making predictions), most recent works have exploited generative models to provide complementary explanations.
no code implementations • 21 Jul 2021 • Haiwen Hong, Xuan Jin, Yin Zhang, Yunqing Hu, Jingfeng Zhang, Yuan He, Hui Xue
In multimodal tasks, we find that the importance of text and image modal information is different for different input cases, and for this motivation, we propose a high-performance and highly general Dual-Router Dynamic Framework (DRDF), consisting of Dual-Router, MWF-Layer, experts and expert fusion unit.
no code implementations • 17 Jul 2021 • Yunqing Hu, Xuan Jin, Yin Zhang, Haiwen Hong, Jingfeng Zhang, Yuan He, Hui Xue
We propose the recurrent attention multi-scale transformer (RAMS-Trans), which uses the transformer's self-attention to recursively learn discriminative region attention in a multi-scale manner.
Ranked #9 on Fine-Grained Image Classification on Stanford Dogs
Fine-Grained Image Classification Fine-Grained Image Recognition
1 code implementation • SIGDIAL (ACL) 2021 • Jingyao Zhou, Haipang Wu, Zehao Lin, Guodun Li, Yin Zhang
Then the representation of each dialogue turn is aggregated by a hierarchical structure to form the passage information, which is utilized in the current turn of DST.
1 code implementation • 8 Jun 2021 • Yueyao Yu, Yin Zhang
To enhance resource efficiency and model deployability of neural networks, we propose a neural-layer architecture based on Householder weighting and absolute-value activating, called Householder-absolute neural layer or simply Han-layer.
no code implementations • 19 May 2021 • Yueyao Yu, Yin Zhang
Despite the tremendous successes of deep neural networks (DNNs) in various applications, many fundamental aspects of deep learning remain incompletely understood, including DNN trainability.
no code implementations • WSDM 2021 • Ziwei Zhu, Yun He, Xing Zhao, Yin Zhang, Jianling Wang, James Caverlee
This paper connects equal opportunity to popularity bias in implicit recommenders to introduce the problem of popularity-opportunity bias.
no code implementations • 1 Jan 2021 • Yueyao Yu, Jie Wang, Wenye Li, Yin Zhang
The stochastic gradient descent (SGD) method, first proposed in 1950's, has been the foundation for deep-neural-network (DNN) training with numerous enhancements including adding a momentum or adaptively selecting learning rates, or using both strategies and more.
no code implementations • COLING 2020 • Qianglong Chen, Feng Ji, Haiqing Chen, Yin Zhang
More concretely, we first introduce a novel graph-based iterative knowledge retrieval module, which iteratively retrieves concepts and entities related to the given question and its choices from multiple knowledge sources.
no code implementations • 29 Oct 2020 • Yin Zhang, Derek Zhiyuan Cheng, Tiansheng Yao, Xinyang Yi, Lichan Hong, Ed H. Chi
It is also very encouraging that our framework further improves head items and overall performance on top of the gains on tail items.
no code implementations • 18 Oct 2020 • Fengda Zhang, Kun Kuang, Zhaoyang You, Tao Shen, Jun Xiao, Yin Zhang, Chao Wu, Yueting Zhuang, Xiaolin Li
FURL poses two new challenges: (1) data distribution shift (Non-IID distribution) among clients would make local models focus on different categories, leading to the inconsistency of representation spaces.
1 code implementation • EMNLP 2020 • Yun He, Zhuoer Wang, Yin Zhang, Ruihong Huang, James Caverlee
We present a new benchmark dataset called PARADE for paraphrase identification that requires specialized domain knowledge.
1 code implementation • EMNLP 2020 • Yun He, Ziwei Zhu, Yin Zhang, Qin Chen, James Caverlee
Knowledge of a disease includes information of various aspects of the disease, such as signs and symptoms, diagnosis and treatment.
1 code implementation • 3 Sep 2020 • Cong Sun, Zhihao Yang, Lei Wang, Yin Zhang, Hongfei Lin, Jian Wang
Using the sequence labeling framework to implement biomedical named entity recognition (BioNER) is currently a conventional method.
no code implementations • 11 Jul 2020 • Renjun Xu, Pelen Liu, Yin Zhang, Fang Cai, Jindong Wang, Shuoying Liang, Heting Ying, Jianwei Yin
However, in a general setting when the target domain contains classes that are never observed in the source domain, namely in Open Set Domain Adaptation (OSDA), existing DA methods failed to work because of the interference of the extra unknown classes.
no code implementations • 27 May 2020 • Zehao Lin, Shaobo Cui, Guodun Li, Xiaoming Kang, Feng Ji, FengLin Li, Zhongzhou Zhao, Haiqing Chen, Yin Zhang
More specifically, we take advantage of a decision model to help the dialogue system decide whether to wait or answer.
no code implementations • 21 May 2020 • Shuke Peng, Feng Ji, Zehao Lin, Shaobo Cui, Haiqing Chen, Yin Zhang
How to build a high-quality multi-domain dialogue system is a challenging work due to its complicated and entangled dialogue state space among each domain, which seriously limits the quality of dialogue policy, and further affects the generated response.
1 code implementation • 19 Mar 2020 • Wenjin Wang, Yunqing Hu, Yin Zhang
To solve those problems, in this paper, we propose a new lifelong learning framework named Searchable Extension Units (SEU) by introducing Neural Architecture Search into lifelong learning, which breaks down the need for a predefined original model and searches for specific extension units for different tasks, without compromising the performance of the model on different tasks.
no code implementations • 22 Feb 2020 • Zehao Lin, Shaobo Cui, Guodun Li, Xiaoming Kang, Feng Ji, FengLin Li, Zhongzhou Zhao, Haiqing Chen, Yin Zhang
And the arbitrator decides whether to wait or to make a response to the user directly.
1 code implementation • 30 Dec 2019 • Yun He, Yin Zhang, Weiwen Liu, James Caverlee
Complementary to methods that exploit specific content patterns (e. g., as in song-based playlists that rely on audio features), the proposed approach models the consistency of item lists based on human curation patterns, and so can be deployed across a wide range of varying item types (e. g., videos, images, books).
1 code implementation • 21 Nov 2019 • Cong Sun, Zhihao Yang, Leilei Su, Lei Wang, Yin Zhang, Hongfei Lin, Jian Wang
Furthermore, the Gaussian probability distribution can effectively improve the extraction performance of sentences with overlapping relations in biomedical relation extraction tasks.
no code implementations • 20 Aug 2019 • Shuke Peng, Xinjing Huang, Zehao Lin, Feng Ji, Haiqing Chen, Yin Zhang
Dialogue systems dealing with multi-domain tasks are highly required.
1 code implementation • 3 Jun 2019 • Feiyu Chen, Yuchen Yang, Liwei Xu, Taiping Zhang, Yin Zhang
The K-means algorithm is arguably the most popular data clustering method, commonly applied to processed datasets in some "feature spaces", as is in spectral clustering.
1 code implementation • Bioinformatics 2019 • Ling Luo, Zhihao Yang, Pei Yang, Yin Zhang, Lei Wang, Hongfei Lin, Jian Wang
Motivation: In biomedical research, chemical is an important class of entities, and chemical named entity recognition (NER) is an important task in the field of biomedical information extraction.
Ranked #6 on Named Entity Recognition (NER) on BC4CHEMD
no code implementations • NeurIPS 2018 • Wenye Li, Jingwei Mao, Yin Zhang, Shuguang Cui
Similarity search is a fundamental problem in computing science with various applications and has attracted significant research attention, especially in large-scale search with high dimensions.
no code implementations • 9 Sep 2018 • Ziwei Zhu, Jianling Wang, Yin Zhang, James Caverlee
This paper highlights our ongoing efforts to create effective information curator recommendation models that can be personalized for individual users, while maintaining important fairness properties.
no code implementations • 6 Mar 2011 • Yangyang Xu, Wotao Yin, Zaiwen Wen, Yin Zhang
By taking the advantages of both nonnegativity and low-rankness, one can generally obtain superior results than those of just using one of the two properties.
Information Theory Numerical Analysis Information Theory Numerical Analysis