no code implementations • 10 Sep 2024 • Atharva Gundawar, Yuchao Li, Dimitri Bertsekas
Structurally, our basic architecture selects moves by a one-move lookahead search, with an intermediate move generated by a nominal opponent engine, and followed by a position evaluation by another chess engine.
no code implementations • 14 Jul 2024 • Ting Bai, Yuchao Li, Karl Henrik Johansson, Jonas Mårtensson
The truck then receives the estimated waiting times from these stations in response, and updates its charging plan accordingly while accounting for travel uncertainties.
no code implementations • 24 May 2024 • Pratyusha Musunuru, Yuchao Li, Jamison Weber, Dimitri Bertsekas
When tested on the MOT17 video dataset, the proposed method demonstrates a 0. 7% improvement in the association accuracy (IDF1 metric) over a state-of-the-art method that is used as the base heuristic.
no code implementations • 19 Mar 2024 • Yuchao Li, Dimitri Bertsekas
We consider methods for computing word sequences that are highly likely, based on these probabilities.
1 code implementation • 12 Nov 2023 • Ting Bai, Yuchao Li, Karl Henrik Johansson, Jonas Mårtensson
Electric trucks usually need to charge their batteries during long-range delivery missions, and the charging times are often nontrivial.
1 code implementation • 19 Sep 2023 • Haojun Xia, Zhen Zheng, Yuchao Li, Donglin Zhuang, Zhongzhu Zhou, Xiafei Qiu, Yong Li, Wei Lin, Shuaiwen Leon Song
Therefore, we propose Flash-LLM for enabling low-cost and highly-efficient large generative model inference with the sophisticated support of unstructured sparsity on high-performance but highly restrictive Tensor Cores.
no code implementations • 15 Mar 2023 • Ting Bai, Yuchao Li, Karl H. Johansson, Jonas Mårtensson
We assume that a collection of charging and rest stations is given along a pre-planned route with known detours and that the problem data are deterministic.
1 code implementation • ICCV 2023 • Song Guo, Lei Zhang, Xiawu Zheng, Yan Wang, Yuchao Li, Fei Chao, Chenglin Wu, Shengchuan Zhang, Rongrong Ji
In this paper, we try to solve this problem by introducing a principled and unified framework based on Information Bottleneck (IB) theory, which further guides us to an automatic pruning approach.
2 code implementations • 23 May 2022 • Yuchao Li, Fuli Luo, Chuanqi Tan, Mengdi Wang, Songfang Huang, Shen Li, Junjie Bai
With the dramatically increased number of parameters in language models, sparsity methods have received ever-increasing research focus to compress and accelerate the models.
1 code implementation • 19 Aug 2021 • Xiawu Zheng, Yuexiao Ma, Teng Xi, Gang Zhang, Errui Ding, Yuchao Li, Jie Chen, Yonghong Tian, Rongrong Ji
This practically limits the application of model compression when the model needs to be deployed on a wide range of devices.
1 code implementation • 4 Jun 2021 • Shaokun Zhang, Xiawu Zheng, Chenyi Yang, Yuchao Li, Yan Wang, Fei Chao, Mengdi Wang, Shen Li, Jun Yang, Rongrong Ji
Motivated by the necessity of efficient inference across various constraints on BERT, we propose a novel approach, YOCO-BERT, to achieve compress once and deploy everywhere.
1 code implementation • 31 May 2021 • Mingbao Lin, Yuxin Zhang, Yuchao Li, Bohong Chen, Fei Chao, Mengdi Wang, Shen Li, Yonghong Tian, Rongrong Ji
We also provide a workflow of filter rearrangement that first rearranges the weight matrix in the output channel dimension to derive more influential blocks for accuracy improvements and then applies similar rearrangement to the next-layer weights in the input channel dimension to ensure correct convolutional operations.
1 code implementation • CVPR 2021 • Yuchao Li, Shaohui Lin, Jianzhuang Liu, Qixiang Ye, Mengdi Wang, Fei Chao, Fan Yang, Jincheng Ma, Qi Tian, Rongrong Ji
Channel pruning and tensor decomposition have received extensive attention in convolutional neural network compression.
1 code implementation • ECCV 2020 • Huixia Li, Chenqian Yan, Shaohui Lin, Xiawu Zheng, Yuchao Li, Baochang Zhang, Fan Yang, Rongrong Ji
Specifically, most state-of-the-art SR models without batch normalization have a large dynamic quantization range, which also serves as another cause of performance drop.
no code implementations • 28 Oct 2020 • Yiwu Yao, Yuchao Li, Chengyu Wang, Tianhang Yu, Houjiang Chen, Xiaotang Jiang, Jun Yang, Jun Huang, Wei Lin, Hui Shu, Chengfei Lv
The intensive computation of Automatic Speech Recognition (ASR) models obstructs them from being deployed on mobile devices.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • L4DC 2020 • Yuchao Li, Karl Henrik Johansson, Jonas Mårtensson
The operator is known to be well-posed for problems with finite states, but our analysis shows that it is also well-defined for the contractive models with infinite states studied.
no code implementations • ECCV 2020 • Yuchao Li, Rongrong Ji, Shaohui Lin, Baochang Zhang, Chenqian Yan, Yongjian Wu, Feiyue Huang, Ling Shao
More specifically, we introduce a novel architecture controlling module in each layer to encode the network architecture by a vector.
1 code implementation • 23 Jan 2019 • Shaohui Lin, Rongrong Ji, Yuchao Li, Cheng Deng, Xuelong. Li
In this paper, we propose a novel filter pruning scheme, termed structured sparsity regularization (SSR), to simultaneously speedup the computation and reduce the memory overhead of CNNs, which can be well supported by various off-the-shelf deep learning libraries.
1 code implementation • CVPR 2019 • Yuchao Li, Shaohui Lin, Baochang Zhang, Jianzhuang Liu, David Doermann, Yongjian Wu, Feiyue Huang, Rongrong Ji
The relationship between the input feature maps and 2D kernels is revealed in a theoretical framework, based on which a kernel sparsity and entropy (KSE) indicator is proposed to quantitate the feature map importance in a feature-agnostic manner to guide model compression.