1 code implementation • 15 Jul 2024 • Zhenxiong Tan, Xinyin Ma, Gongfan Fang, Xinchao Wang
Latent diffusion models have shown promising results in audio generation, making notable advancements over traditional methods.
1 code implementation • 5 Jul 2024 • Gongfan Fang, Xinyin Ma, Michael Bi Mi, Xinchao Wang
For instance, we improve the accuracy of DeiT-Tiny from 74. 52% to 77. 50% by pruning an off-the-shelf DeiT-Base model.
2 code implementations • 11 Jun 2024 • Zigeng Chen, Xinyin Ma, Gongfan Fang, Zhenxiong Tan, Xinchao Wang
To address this, we introduce AsyncDiff, a universal and plug-and-play acceleration scheme that enables model parallelism across multiple devices.
1 code implementation • 3 Jun 2024 • Xinyin Ma, Gongfan Fang, Michael Bi Mi, Xinchao Wang
To address the challenge of the exponential search space in deep models for identifying layers to cache and remove, we propose a novel differentiable optimization objective.
2 code implementations • 8 Dec 2023 • Zigeng Chen, Gongfan Fang, Xinyin Ma, Xinchao Wang
To address this challenging trade-off, we introduce SlimSAM, a novel data-efficient SAM compression method that achieves superior performance with extremely less training data.
2 code implementations • CVPR 2024 • Xinyin Ma, Gongfan Fang, Xinchao Wang
Diffusion models have recently gained unprecedented attention in the field of image synthesis due to their remarkable generative capabilities.
2 code implementations • NeurIPS 2023 • Xinyin Ma, Gongfan Fang, Xinchao Wang
With LLM being a general-purpose task solver, we explore its compression in a task-agnostic manner, which aims to preserve the multi-task solving and language generation ability of the original LLM.
1 code implementation • NeurIPS 2023 • Gongfan Fang, Xinyin Ma, Xinchao Wang
Generative modeling has recently undergone remarkable advancements, primarily propelled by the transformative implications of Diffusion Probabilistic Models (DPMs).
1 code implementation • CVPR 2023 • Gongfan Fang, Xinyin Ma, Mingli Song, Michael Bi Mi, Xinchao Wang
Structural pruning enables model acceleration by removing structurally-grouped parameters from neural networks.
no code implementations • 16 May 2022 • Xinyin Ma, Xinchao Wang, Gongfan Fang, Yongliang Shen, Weiming Lu
Data-free knowledge distillation (DFKD) conducts knowledge distillation via eliminating the dependence of original training data, and has recently achieved impressive results in accelerating pre-trained language models.
1 code implementation • EMNLP 2021 • Xinyin Ma, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Weiming Lu
Entity retrieval, which aims at disambiguating mentions to canonical entities from massive KBs, is essential for many tasks in natural language processing.
Ranked #1 on Entity Retrieval on ZESHEL
1 code implementation • ACL 2021 • Yongliang Shen, Xinyin Ma, Zeqi Tan, Shuai Zhang, Wen Wang, Weiming Lu
Although these methods have the innate ability to handle nested NER, they suffer from high computational cost, ignorance of boundary information, under-utilization of the spans that partially match with entities, and difficulties in long entity recognition.
Ranked #6 on Nested Named Entity Recognition on GENIA
Chinese Named Entity Recognition named-entity-recognition +3
1 code implementation • 25 Jan 2021 • Yongliang Shen, Xinyin Ma, Yechun Tang, Weiming Lu
Joint entity and relation extraction framework constructs a unified model to perform entity recognition and relation extraction simultaneously, which can exploit the dependency between the two tasks to mitigate the error propagation problem suffered by the pipeline model.
Ranked #1 on Relation Extraction on CoNLL04 (NER Micro F1 metric)
Joint Entity and Relation Extraction Reading Comprehension +2
no code implementations • Findings of the Association for Computational Linguistics 2020 • Jiale Yu, Yongliang Shen, Xinyin Ma, Chenghao Jia, Chen Chen, Weiming Lu
Extensive experiments on a real-world dataset show the effectiveness of our approach.
no code implementations • EMNLP 2020 • Xinyin Ma, Yongliang Shen, Gongfan Fang, Chen Chen, Chenghao Jia, Weiming Lu
To the best of our knowledge, our framework is the first data-free distillation framework designed for NLP tasks.
no code implementations • 11 Jun 2020 • Zeyun Tang, Yongliang Shen, Xinyin Ma, Wei Xu, Jiale Yu, Weiming Lu
Meanwhile, we propose Gated-RGCN to accumulate evidence on the path-based reasoning graph, which contains a new question-aware gating mechanism to regulate the usefulness of information propagating across documents and add question information during reasoning.