no code implementations • 10 Sep 2024 • Linfeng Zhang, Changyue Hu, Zhiyu Quan
As the body of academic literature continues to grow, researchers face increasing difficulties in effectively searching for relevant resources.
1 code implementation • 28 Aug 2024 • Sihang Li, Jin Huang, Jiaxi Zhuang, Yaorui Shi, Xiaochen Cai, Mingjun Xu, Xiang Wang, Linfeng Zhang, Guolin Ke, Hengxing Cai
To develop an LLM specialized in scientific literature understanding, we propose a hybrid strategy that integrates continual pre-training (CPT) and supervised fine-tuning (SFT), to simultaneously infuse scientific domain knowledge and enhance instruction-following capabilities for domain-specific tasks. cIn this process, we identify two key challenges: (1) constructing high-quality CPT corpora, and (2) generating diverse SFT instructions.
no code implementations • 22 Aug 2024 • Shaobo Wang, Yantai Yang, Qilong Wang, Kaixin Li, Linfeng Zhang, Junchi Yan
Our findings suggest that prioritizing the synthesis of easier samples from the original dataset can enhance the quality of distilled datasets, especially in low IPC (image-per-class) settings.
1 code implementation • 8 Jul 2024 • Zhanghao Zhouyin, Zixi Gan, Shishir Kumar Pandey, Linfeng Zhang, Qiangqiang Gu
Predicting quantum operator matrices such as Hamiltonian, overlap, and density matrices in the density functional theory (DFT) framework is crucial for understanding material properties.
no code implementations • 8 Jul 2024 • Boshen Zeng, SiAn Chen, Xinxin Liu, Changhong Chen, Bin Deng, Xiaoxu Wang, Zhifeng Gao, Yuzhi Zhang, Weinan E, Linfeng Zhang
Advancements in lithium battery technology heavily rely on the design and engineering of electrolytes.
1 code implementation • 21 Jun 2024 • Xiaohong Ji, Zhen Wang, Zhifeng Gao, Hang Zheng, Linfeng Zhang, Guolin Ke, Weinan E
In recent years, pretraining models have made significant advancements in the fields of natural language processing (NLP), computer vision (CV), and life sciences.
no code implementations • 12 Jun 2024 • Zhihang Yuan, Pu Lu, Hanling Zhang, Xuefei Ning, Linfeng Zhang, Tianchen Zhao, Shengen Yan, Guohao Dai, Yu Wang
We identify three key redundancies in the attention computation during DiT inference: 1. spatial redundancy, where many attention heads focus on local information; 2. temporal redundancy, with high similarity between neighboring steps' attention outputs; 3. conditional redundancy, where conditional and unconditional inferences exhibit significant similarity.
1 code implementation • 20 May 2024 • Eric Alcaide, Zhifeng Gao, Guolin Ke, Yaqi Li, Linfeng Zhang, Hang Zheng, Gengmo Zhou
In recent years, machine learning (ML) methods have emerged as promising alternatives for molecular docking, offering the potential for high accuracy without incurring prohibitive computational costs.
no code implementations • 15 Mar 2024 • Hengxing Cai, Xiaochen Cai, Shuwen Yang, Jiankun Wang, Lin Yao, Zhifeng Gao, Junhan Chang, Sihang Li, Mingjun Xu, Changxin Wang, Hongshuai Wang, Yongge Li, Mujie Lin, Yaqi Li, Yuqi Yin, Linfeng Zhang, Guolin Ke
Scientific literature often includes a wide range of multimodal elements, such as tables, charts, and molecule, which are hard for text-focused LLMs to understand and analyze.
no code implementations • 4 Mar 2024 • Hengxing Cai, Xiaochen Cai, Junhan Chang, Sihang Li, Lin Yao, Changxin Wang, Zhifeng Gao, Hongshuai Wang, Yongge Li, Mujie Lin, Shuwen Yang, Jiankun Wang, Mingjun Xu, Jin Huang, Fang Xi, Jiaxi Zhuang, Yuqi Yin, Yaqi Li, Changhong Chen, Zheng Cheng, Zifeng Zhao, Linfeng Zhang, Guolin Ke
Recent breakthroughs in Large Language Models (LLMs) have revolutionized natural language understanding and generation, sparking significant interest in applying them to scientific literature analysis.
no code implementations • 2 Aug 2023 • Wing Fung Chong, Daniel Linders, Zhiyu Quan, Linfeng Zhang
In the current market practice, many cyber insurance products offer a coverage bundle for losses arising from various types of incidents, such as data breaches and ransomware attacks, and the coverage for each incident type comes with a separate limit and deductible.
no code implementations • ICCV 2023 • Tianchen Zhao, Xuefei Ning, Ke Hong, Zhongyuan Qiu, Pu Lu, Yali Zhao, Linfeng Zhang, Lipu Zhou, Guohao Dai, Huazhong Yang, Yu Wang
One reason for this high resource consumption is the presence of a large number of redundant background points in Lidar point clouds, resulting in spatial redundancy in both 3D voxel and dense BEV map representations.
no code implementations • 22 May 2023 • Muzhou Yu, Linfeng Zhang, Kaisheng Ma
In this paper, we revisit the usage of data augmentation in model compression and give a comprehensive study on the relation between model sizes and their optimal data augmentation policy.
no code implementations • 28 Apr 2023 • Muzhou Yu, Sia Huat Tan, Kailu Wu, Runpei Dong, Linfeng Zhang, Kaisheng Ma
Knowledge distillation conducts an effective model compression method while holding some limitations:(1) the feature based distillation methods only focus on distilling the feature map but are lack of transferring the relation of data examples; (2) the relational distillation methods are either limited to the handcrafted functions for relation extraction, such as L2 norm, or weak in inter- and intra- class relation modeling.
1 code implementation • 24 Apr 2023 • Zhifeng Gao, Xiaohong Ji, Guojiang Zhao, Hongshuai Wang, Hang Zheng, Guolin Ke, Linfeng Zhang
Recently deep learning based quantitative structure-activity relationship (QSAR) models has shown surpassing performance than traditional methods for property prediction tasks in drug discovery.
1 code implementation • 16 Mar 2023 • Shuqi Lu, Zhifeng Gao, Di He, Linfeng Zhang, Guolin Ke
Uni-Mol+ first generates a raw 3D molecule conformation from inexpensive methods such as RDKit.
no code implementations • ICCV 2023 • Fei Li, Linfeng Zhang, Zikun Liu, Juan Lei, Zhenbo Li
CNN's limited receptive field restricts its ability to capture long-range spatial-temporal dependencies, leading to unsatisfactory performance in video super-resolution.
no code implementations • ICCV 2023 • Linfeng Zhang, Kaisheng Ma
Significant advancements have been accomplished with deep neural networks in diverse visual tasks, which have substantially elevated their deployment in edge device software.
4 code implementations • 16 Dec 2022 • Runpei Dong, Zekun Qi, Linfeng Zhang, Junbo Zhang, Jianjian Sun, Zheng Ge, Li Yi, Kaisheng Ma
The success of deep learning heavily relies on large-scale data with comprehensive labels, which is more expensive and time-consuming to fetch in 3D compared to 2D images or natural languages.
Ranked #6 on Few-Shot 3D Point Cloud Classification on ModelNet40 10-way (10-shot) (using extra training data)
Few-Shot 3D Point Cloud Classification Knowledge Distillation +1
no code implementations • 14 Nov 2022 • Linfeng Zhang, Yukang Shi, Hung-Shuo Tai, Zhipeng Zhang, Yuan He, Ke Wang, Kaisheng Ma
Detecting 3D objects from multi-view images is a fundamental problem in 3D computer vision.
1 code implementation • 13 Sep 2022 • Hao Xie, Zi-Hang Li, Han Wang, Linfeng Zhang, Lei Wang
We developed a deep generative model-based variational free energy approach to the equations of state of dense hydrogen.
1 code implementation • ChemRxiv 2022 • Gengmo Zhou, Zhifeng Gao, Qiankun Ding, Hang Zheng, Hongteng Xu, Zhewei Wei, Linfeng Zhang, Guolin Ke
Uni-Mol is composed of two models with the same SE(3)-equivariant transformer architecture: a molecular pretraining model trained by 209M molecular conformations; a pocket pretraining model trained by 3M candidate protein pocket data.
Ranked #1 on Molecular Property Prediction on MUV
1 code implementation • 17 Aug 2022 • Duo Zhang, Hangrui Bi, Fu-Zhi Dai, Wanrun Jiang, Linfeng Zhang, Han Wang
Machine learning assisted modeling of the inter-atomic potential energy surface (PES) is revolutionizing the field of molecular simulation.
1 code implementation • 12 Jul 2022 • Linfeng Zhang, Xin Chen, Junbo Zhang, Runpei Dong, Kaisheng Ma
The success of deep learning is usually accompanied by the growth in neural network depth.
1 code implementation • 21 Jun 2022 • Wenfei Li, Qi Ou, Yixiao Chen, Yu Cao, Renxi Liu, Chunyi Zhang, Daye Zheng, Chun Cai, Xifan Wu, Han Wang, Mohan Chen, Linfeng Zhang
However, for high-level QM methods, such as density functional theory (DFT) at the meta-GGA level and/or with exact exchange, quantum Monte Carlo, etc., generating a sufficient amount of data for training a ML potential has remained computationally challenging due to their high cost.
no code implementations • 25 May 2022 • Linfeng Zhang, Xin Chen, Runpei Dong, Kaisheng Ma
In this paper, we propose Region-aware Knowledge Distillation ReKo to compress image-to-image translation models.
1 code implementation • CVPR 2023 • Linfeng Zhang, Runpei Dong, Hung-Shuo Tai, Kaisheng Ma
The remarkable breakthroughs in point cloud representation learning have boosted their usage in real-world applications such as self-driving cars and virtual reality.
no code implementations • 17 May 2022 • Wing Fung Chong, Runhuan Feng, Hins Hu, Linfeng Zhang
This framework is demonstrated by a case study based on a historical cyber incident dataset, which shows that a comprehensive cost-benefit analysis is necessary for a budget-constrained company with competing objectives for cyber risk management.
no code implementations • CVPR 2022 • Linfeng Zhang, Xin Chen, Xiaobing Tu, Pengfei Wan, Ning Xu, Kaisheng Ma
Instead of directly distilling the generated images of teachers, wavelet knowledge distillation first decomposes the images into different frequency bands with discrete wavelet transformation and then only distills the high frequency bands.
1 code implementation • 10 Jan 2022 • Hao Xie, Linfeng Zhang, Lei Wang
The quasiparticle effective mass $m^\ast$ of interacting electrons is a fundamental quantity in the Fermi liquid theory.
1 code implementation • 30 Dec 2021 • Runpei Dong, Zhanhong Tan, Mengdi Wu, Linfeng Zhang, Kaisheng Ma
Besides, an efficient deployment flow for the mobile CPU is developed, achieving up to 7. 46$\times$ inference acceleration on an octa-core ARM CPU.
no code implementations • 29 Sep 2021 • Linfeng Zhang, Kaisheng Ma
To tackle this challenge, in this paper, we propose Region-aware Knowledge Distillation which first localizes the crucial regions in the images with attention mechanism.
1 code implementation • 18 May 2021 • Hao Xie, Linfeng Zhang, Lei Wang
The variational density matrix is parametrized by a permutation equivariant many-body unitary transformation together with a discrete probabilistic model.
no code implementations • 9 Feb 2021 • Linfeng Zhang, Han Wang, Roberto Car, Weinan E
Using the Deep Potential methodology, we construct a model that reproduces accurately the potential energy surface of the SCAN approximation of density functional theory for water, from low temperature and pressure to about 2400 K and 50 GPa, excluding the vapor stability region.
Chemical Physics
1 code implementation • ICLR 2021 • Linfeng Zhang, Kaisheng Ma
In this paper, we suggest that the failure of knowledge distillation on object detection is mainly caused by two reasons: (1) the imbalance between pixels of foreground and background and (2) lack of distillation on the relation between different pixels.
no code implementations • 9 Dec 2020 • Jianhang Xu, Chunyi Zhang, Linfeng Zhang, Mohan Chen, Biswajit Santra, Xifan Wu
Feynman path-integral deep potential molecular dynamics (PI-DPMD) calculations have been employed to study both light (H$_2$O) and heavy water (D$_2$O) within the isothermal-isobaric ensemble.
Chemical Physics Computational Physics
no code implementations • 6 Dec 2020 • Xiaowei Chen, Wing Fung Chong, Runhuan Feng, Linfeng Zhang
Repeated history of pandemics, such as SARS, H1N1, Ebola, Zika, and COVID-19, has shown that pandemic risk is inevitable.
1 code implementation • NeurIPS 2020 • Linfeng Zhang, Yukang Shi, Zuoqiang Shi, Kaisheng Ma, Chenglong Bao
Moreover, an orthogonal loss is applied to the feature resizing layer in TOFD to improve the performance of knowledge distillation.
no code implementations • 1 Aug 2020 • Yixiao Chen, Linfeng Zhang, Han Wang, E Weinan
We propose a general machine learning-based framework for building an accurate and widely-applicable energy functional within the framework of generalized Kohn-Sham density functional theory.
no code implementations • 5 Jun 2020 • Jianxing Huang, Linfeng Zhang, Han Wang, Jinbao Zhao, Jun Cheng, Weinan E
It has been a challenge to accurately simulate Li-ion diffusion processes in battery materials at room temperature using {\it ab initio} molecular dynamics (AIMD) due to its high computational cost.
Computational Physics Materials Science Chemical Physics
no code implementations • 4 Jun 2020 • Weinan E, Jiequn Han, Linfeng Zhang
Machine learning is poised as a very powerful tool that can drastically improve our ability to carry out scientific research.
no code implementations • CVPR 2020 • Linfeng Zhang, Muzhou Yu, Tong Chen, Zuoqiang Shi, Chenglong Bao, Kaisheng Ma
In the training stage, a novel distillation method named input-aware self distillation is proposed to facilitate the primary classifier to learn the robust information from auxiliary classifiers.
1 code implementation • 1 May 2020 • Weile Jia, Han Wang, Mohan Chen, Denghui Lu, Lin Lin, Roberto Car, Weinan E, Linfeng Zhang
For 35 years, {\it ab initio} molecular dynamics (AIMD) has been the method of choice for modeling complex atomistic phenomena from first principles.
Computational Physics
1 code implementation • 5 Dec 2019 • Yuni Lai, Linfeng Zhang, Donghong Han, Rui Zhou, Guoren Wang
In addition, a pooling method based on percentile is proposed to improve the accuracy of the model.
no code implementations • 27 Nov 2019 • Leonardo Zepeda-Núñez, Yixiao Chen, Jiefu Zhang, Weile Jia, Linfeng Zhang, Lin Lin
By directly targeting at the self-consistent electron density, we demonstrate that the adapted network architecture, called the Deep Density, can effectively represent the electron density as the linear combination of contributions from many local clusters.
1 code implementation • 28 Oct 2019 • Yuzhi Zhang, Haidi Wang, WeiJie Chen, Jinzhe Zeng, Linfeng Zhang, Han Wang, Weinan E
Materials 3, 023804] and is capable of generating uniformly accurate deep learning based PES models in a way that minimizes human intervention and the computational cost for data generation and model training.
Computational Physics
1 code implementation • 30 Sep 2019 • Shuo-Hui Li, Chen-Xiao Dong, Linfeng Zhang, Lei Wang
We construct flexible and powerful canonical transformations as generative models using symplectic neural networks.
no code implementations • 3 Jul 2019 • Xiaolong Ma, Sheng Lin, Shaokai Ye, Zhezhi He, Linfeng Zhang, Geng Yuan, Sia Huat Tan, Zhengang Li, Deliang Fan, Xuehai Qian, Xue Lin, Kaisheng Ma, Yanzhi Wang
Based on the proposed comparison framework, with the same accuracy and quantization, the results show that non-structrued pruning is not competitive in terms of both storage and computation efficiency.
1 code implementation • 27 Jun 2019 • Linfeng Zhang, Mohan Chen, Xifan Wu, Han Wang, Weinan E, Roberto Car
We introduce a deep neural network (DNN) model that assigns the position of the centers of the electronic charge in each atomic configuration on a molecular dynamics trajectory.
Computational Physics Materials Science Chemical Physics
1 code implementation • NeurIPS 2019 • Linfeng Zhang, Zhanhong Tan, Jiebo Song, Jingwei Chen, Chenglong Bao, Kaisheng Ma
Remarkable achievements have been attained by deep neural networks in various applications.
1 code implementation • ICCV 2019 • Linfeng Zhang, Jiebo Song, Anni Gao, Jingwei Chen, Chenglong Bao, Kaisheng Ma
Different from traditional knowledge distillation - a knowledge transformation methodology among networks, which forces student neural networks to approximate the softmax layer outputs of pre-trained teacher neural networks, the proposed self distillation framework distills knowledge within network itself.
no code implementations • ICLR 2019 • Linfeng Zhang, Weinan E, Lei Wang
We present a deep generative model, named Monge-Amp\`ere flow, which builds on continuous-time gradient flow arising from the Monge-Amp\`ere equation in optimal transport theory.
no code implementations • 28 Oct 2018 • Linfeng Zhang, De-Ye Lin, Han Wang, Roberto Car, Weinan E
An active learning procedure called Deep Potential Generator (DP-GEN) is proposed for the construction of accurate and transferable machine learning-based models of the potential energy surface (PES) for the molecular modeling of materials.
1 code implementation • 26 Sep 2018 • Linfeng Zhang, Weinan E, Lei Wang
We present a deep generative model, named Monge-Amp\`ere flow, which builds on continuous-time gradient flow arising from the Monge-Amp\`ere equation in optimal transport theory.
1 code implementation • 29 Jul 2018 • Tianyun Zhang, Shaokai Ye, Kaiqi Zhang, Xiaolong Ma, Ning Liu, Linfeng Zhang, Jian Tang, Kaisheng Ma, Xue Lin, Makan Fardad, Yanzhi Wang
Without loss of accuracy on the AlexNet model, we achieve 2. 58X and 3. 65X average measured speedup on two GPUs, clearly outperforming the prior work.
no code implementations • 18 Jul 2018 • Jiequn Han, Linfeng Zhang, Weinan E
We introduce a new family of trial wave-functions based on deep neural networks to solve the many-electron Schr\"odinger equation.
Computational Physics Chemical Physics
1 code implementation • NeurIPS 2018 • Linfeng Zhang, Jiequn Han, Han Wang, Wissam A. Saidi, Roberto Car, Weinan E
Machine learning models are changing the paradigm of molecular modeling, which is a fundamental tool for material science, chemistry, and computational biology.
Computational Physics Materials Science Chemical Physics
2 code implementations • 11 Dec 2017 • Han Wang, Linfeng Zhang, Jiequn Han, Weinan E
Here we describe DeePMD-kit, a package written in Python/C++ that has been designed to minimize the effort required to build deep learning based representation of potential energy and force field and to perform molecular dynamics.
no code implementations • 10 Dec 2017 • Linfeng Zhang, Han Wang, Weinan E
Like metadynamics, it allows for an efficient exploration of the configuration space by adding an adaptively computed biasing potential to the original dynamics.
5 code implementations • 30 Jul 2017 • Linfeng Zhang, Jiequn Han, Han Wang, Roberto Car, Weinan E
We introduce a scheme for molecular simulations, the Deep Potential Molecular Dynamics (DeePMD) method, based on a many-body potential and interatomic forces generated by a carefully crafted deep neural network trained with ab initio data.
1 code implementation • 5 Jul 2017 • Jiequn Han, Linfeng Zhang, Roberto Car, Weinan E
When tested on a wide variety of examples, Deep Potential is able to reproduce the original model, whether empirical or quantum mechanics based, within chemical accuracy.
Computational Physics