Search Results for author: Jue Chen

Found 4 papers, 2 papers with code

Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement

1 code implementation1 Nov 2024 Yingwei Ma, Rongyu Cao, Yongchang Cao, Yue Zhang, Jue Chen, Yibo Liu, Yuchen Liu, Binhua Li, Fei Huang, Yongbin Li

The results demonstrate that Lingma SWE-GPT 72B successfully resolves 30. 20% of the GitHub issues, marking a significant improvement in automatic issue resolution (22. 76% relative improvement compared to Llama 3. 1 405B), approaching the performance of closed-source models (31. 80\% issues of GPT-4o resolved).

Language Modeling Language Modelling

Resource Constrained Model Compression via Minimax Optimization for Spiking Neural Networks

1 code implementation9 Aug 2023 Jue Chen, Huan Yuan, Jianchao Tan, Bin Chen, Chengru Song, Di Zhang

We propose an improved end-to-end Minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency.

Model Compression Sparse Learning

Razor SNN: Efficient Spiking Neural Network with Temporal Embeddings

no code implementations30 Jun 2023 Yuan Zhang, Jian Cao, Ling Zhang, Jue Chen, Wenyu Sun, YuAn Wang

The event streams generated by dynamic vision sensors (DVS) are sparse and non-uniform in the spatial domain, while still dense and redundant in the temporal domain.

Boosting Pruned Networks with Linear Over-parameterization

no code implementations25 Apr 2022 Yu Qian, Jian Cao, Xiaoshuang Li, Jie Zhang, Hufei Li, Jue Chen

To address this challenge, we propose a novel method that first linearly over-parameterizes the compact layers in pruned networks to enlarge the number of fine-tuning parameters and then re-parameterizes them to the original layers after fine-tuning.

Knowledge Distillation

Cannot find the paper you are looking for? You can Submit a new open access paper.