1 code implementation • 23 Dec 2024 • Linhao Zhang, Daoguang Zan, Quanshun Yang, Zhirong Huang, Dong Chen, Bo Shen, Tianyu Liu, Yongshun Gong, Pengjie Huang, Xudong Lu, Guangtai Liang, Lizhen Cui, Qianxiang Wang
Large Language Models (LLMs) have advanced rapidly in recent years, with their applications in software engineering expanding to more complex repository-level tasks.
no code implementations • 30 Sep 2024 • Haiyan Zhao, Heng Zhao, Bo Shen, Ali Payani, Fan Yang, Mengnan Du
Training linear classifiers on probing tasks is a principle approach to denote the vector of a certain concept in the representation space.
no code implementations • 7 Sep 2024 • Bo Shen, Marco Marena, Chenyang Li, Qin Li, Haodi Jiang, Mengnan Du, Jiajun Xu, Haimin Wang
With recent missions such as advanced space-based observatories like the Solar Dynamics Observatory (SDO) and Parker Solar Probe, and ground-based telescopes like the Daniel K. Inouye Solar Telescope (DKIST), the volume, velocity, and variety of data have made solar physics enter a transformative era as solar physics big data (SPBD).
2 code implementations • 26 Aug 2024 • Daoguang Zan, Zhirong Huang, Ailun Yu, Shaoxin Lin, Yifan Shi, Wei Liu, Dong Chen, Zongshuai Qi, Hao Yu, Lei Yu, Dezhi Ran, Muhan Zeng, Bo Shen, Pan Bian, Guangtai Liang, Bei guan, Pengjie Huang, Tao Xie, Yongji Wang, Qianxiang Wang
GitHub issue resolving is a critical task in software engineering, recently gaining significant attention in both industry and academia.
1 code implementation • 21 May 2024 • Yutao Du, Qin Li, Raghav Gnanasambandam, Mengnan Du, Haimin Wang, Bo Shen
The goal of this study is to accelerate coronal magnetic field simulation using deep learning, specifically, the Fourier Neural Operator (FNO).
2 code implementations • 25 Mar 2024 • Daoguang Zan, Ailun Yu, Wei Liu, Dong Chen, Bo Shen, Wei Li, Yafen Yao, Yongshun Gong, Xiaolin Chen, Bei guan, Zhiguang Yang, Yongji Wang, Qianxiang Wang, Lizhen Cui
For feedback-based evaluation, we develop a VSCode plugin for CodeS and engage 30 participants in conducting empirical studies.
no code implementations • 16 Feb 2024 • Haiyan Zhao, Fan Yang, Bo Shen, Himabindu Lakkaraju, Mengnan Du
Large language models (LLMs) have led to breakthroughs in language tasks, yet the internal mechanisms that enable their remarkable generalization and reasoning abilities remain opaque.
no code implementations • 31 Aug 2023 • Daoguang Zan, Ailun Yu, Bo Shen, Jiaxin Zhang, Taihong Chen, Bing Geng, Bei Chen, Jichuan Ji, Yafen Yao, Yongji Wang, Qianxiang Wang
Results demonstrate that programming languages can significantly improve each other.
no code implementations • 27 Jul 2023 • Bo Shen, Jiaxin Zhang, Taihong Chen, Daoguang Zan, Bing Geng, An Fu, Muhan Zeng, Ailun Yu, Jichuan Ji, Jingyang Zhao, Yuenan Guo, Qianxiang Wang
In this paper, we propose a novel RRTF (Rank Responses to align Test&Teacher Feedback) framework, which can effectively and efficiently boost pre-trained large language models for code generation.
1 code implementation • 13 Mar 2023 • Chenyang Li, Jihoon Chung, Mengnan Du, Haimin Wang, Xianlian Zhou, Bo Shen
This paper focuses on two model compression techniques: low-rank approximation and weight pruning in neural networks, which are very popular nowadays.
no code implementations • 19 Nov 2022 • Lifu Wang, Tianyu Wang, Shengwei Yi, Bo Shen, Bo Hu, Xing Cao
We study the learning ability of linear recurrent neural networks with Gradient Descent.
no code implementations • 28 Oct 2022 • Jihoon Chung, Bo Shen, Andrew Chung Chee Law, Zhenyu, Kong
Since AM typically fabricates a small number of customized products, this paper aims to create an online learning-based strategy to mitigate the new defects in AM process while minimizing the number of samples needed.
no code implementations • 28 Oct 2022 • Jihoon Chung, Bo Shen, Zhenyu, Kong
It is beneficial to generate effective artificial sample data for the abnormal states to make a more balanced training set.
no code implementations • 28 Oct 2022 • Jihoon Chung, Bo Shen, Zhenyu, Kong
Fault diagnosis is to identify process faults that cause the excessive dimensional variation of the product using dimensional measurements.
1 code implementation • 22 Jul 2022 • Fenia Christopoulou, Gerasimos Lampouras, Milan Gritta, Guchun Zhang, Yinpeng Guo, Zhongqi Li, Qi Zhang, Meng Xiao, Bo Shen, Lin Li, Hao Yu, Li Yan, Pingyi Zhou, Xin Wang, Yuchi Ma, Ignacio Iacobacci, Yasheng Wang, Guangtai Liang, Jiansheng Wei, Xin Jiang, Qianxiang Wang, Qun Liu
We present PanGu-Coder, a pretrained decoder-only language model adopting the PanGu-Alpha architecture for text-to-code generation, i. e. the synthesis of programming language solutions given a natural language problem description.
no code implementations • 26 Apr 2022 • Raghav Gnanasambandam, Bo Shen, Jihoon Chung, Xubo Yue, Zhenyu, Kong
To address this, a Self-scalable tanh (Stan) activation function is proposed for the PINNs.
no code implementations • 29 Mar 2022 • Bo Shen, Weijun Xie, Zhenyu Kong
The objective of this study is to address the problem of background/foreground separation with missing pixels by combining the video acquisition, video recovery, background/foreground separation into a single framework.
no code implementations • NeurIPS 2021 • Lifu Wang, Bo Shen, Bo Hu, Xing Cao
In this paper, using detailed analysis about the neural tangent kernel matrix, we prove a generalization error bound to learn such functions without normalized conditions and show that some notable concept classes are learnable with the numbers of iterations and samples scaling almost-polynomially in the input length $L$.
no code implementations • 5 Aug 2020 • Bo Shen, Zhenyu, Kong
Tensor, also known as multi-dimensional array, arises from many applications in signal processing, manufacturing processes, healthcare, among others.
no code implementations • 10 Jun 2020 • Lifu Wang, Bo Shen, Ning Zhao, Zhiyuan Zhang
In this paper, we follow this line to study the topology (sub-level sets) of the loss landscape of deep ReLU neural networks with a skip connection and theoretically prove that the skip connection network inherits the good properties of the two-layer network and skip connections can help to control the connectedness of the sub-level sets, such that any local minima worse than the global minima of some two-layer ReLU network will be very ``shallow".
no code implementations • 14 Oct 2019 • Lifu Wang, Bo Shen, Ning Zhao
This may have a negative influence on the convergence of the algorithm.
no code implementations • 28 Nov 2018 • Bo Shen, Wei zhang, Haiyan Zhao, Zhi Jin, Yanhong Wu
And through feedback, each player is provided with personalized feedback information based on the current COG and the player's exploration result, in order to accelerate his/her puzzle-solving process.