Search Results for author: Jishen Zhao

Found 17 papers, 5 papers with code

Multi-modal Learning for WebAssembly Reverse Engineering

no code implementations4 Apr 2024 Hanxian Huang, Jishen Zhao

WasmRev is pre-trained using self-supervised learning on a large-scale multi-modal corpus encompassing source code, code documentation and the compiled WebAssembly, without requiring labeled data.

Language Modelling Self-Supervised Learning

GeoT: Tensor Centric Library for Graph Neural Network via Efficient Segment Reduction on GPU

1 code implementation3 Apr 2024 Zhongming Yu, Genghan Zhang, Hanxian Huang, Xin Chen, Jishen Zhao

Yet, efficient tensor-centric frameworks for GNNs remain scarce due to unique challenges and limitations encountered when implementing segment reduction in GNN contexts.

Learning to Maximize Mutual Information for Chain-of-Thought Distillation

no code implementations5 Mar 2024 Xin Chen, Hanxian Huang, Yanjun Gao, Yi Wang, Jishen Zhao, Ke Ding

Knowledge distillation, the technique of transferring knowledge from large, complex models to smaller ones, marks a pivotal step towards efficient AI deployment.

Knowledge Distillation Language Modelling +1

Sibyl: Forecasting Time-Evolving Query Workloads

no code implementations8 Jan 2024 Hanxian Huang, Tarique Siddiqui, Rana Alotaibi, Carlo Curino, Jyoti Leeka, Alekh Jindal, Jishen Zhao, Jesus Camacho-Rodriguez, Yuanyuan Tian

Drawing insights from real-workloads, we propose template-based featurization techniques and develop a stacked-LSTM with an encoder-decoder architecture for accurate forecasting of query workloads.

Interpretable and Flexible Target-Conditioned Neural Planners For Autonomous Vehicles

no code implementations23 Sep 2023 Haolan Liu, Jishen Zhao, Liangjun Zhang

Learning-based approaches to autonomous vehicle planners have the potential to scale to many complicated real-world driving scenarios by leveraging huge amounts of driver demonstrations.

Autonomous Vehicles Collision Avoidance

TripLe: Revisiting Pretrained Model Reuse and Progressive Learning for Efficient Vision Transformer Scaling and Searching

no code implementations ICCV 2023 Cheng Fu, Hanxian Huang, Zixuan Jiang, Yun Ni, Lifeng Nai, Gang Wu, Liqun Cheng, Yanqi Zhou, Sheng Li, Andrew Li, Jishen Zhao

One promising way to accelerate transformer training is to reuse small pretrained models to initialize the transformer, as their existing representation power facilitates faster model convergence.

Knowledge Distillation Neural Architecture Search

Everyone's Preference Changes Differently: Weighted Multi-Interest Retrieval Model

1 code implementation14 Jul 2022 Hui Shi, Yupeng Gu, Yitong Zhou, Bo Zhao, Sicun Gao, Jishen Zhao

In this paper, we propose the Multi-Interest Preference (MIP) model, an approach that not only produces multi-interest for users by using the user's sequential engagement more effectively but also automatically learns a set of weights to represent the preference over each embedding so that the candidates can be retrieved from each interest proportionally.

Recommendation Systems Retrieval

Learning Bounded Context-Free-Grammar via LSTM and the Transformer:Difference and Explanations

1 code implementation16 Dec 2021 Hui Shi, Sicun Gao, Yuandong Tian, Xinyun Chen, Jishen Zhao

With the forced decomposition, we show that the performance upper bounds of LSTM and Transformer in learning CFL are close: both of them can simulate a stack and perform stack operation along with state transitions.

ProFlip: Targeted Trojan Attack With Progressive Bit Flips

no code implementations ICCV 2021 Huili Chen, Cheng Fu, Jishen Zhao, Farinaz Koushanfar

In this work, we present ProFlip, the first targeted Trojan attack framework that can divert the prediction of the DNN to the target class by progressively identifying and flipping a small set of bits in model parameters.

Continuous Convolutional Neural Network forNonuniform Time Series

no code implementations25 Sep 2019 Hui Shi, Yang Zhang, Hao Wu, Shiyu Chang, Kaizhi Qian, Mark Hasegawa-Johnson, Jishen Zhao

Convolutional neural network (CNN) for time series data implicitly assumes that the data are uniformly sampled, whereas many event-based and multi-modal data are nonuniform or have heterogeneous sampling rates.

Time Series Time Series Analysis

A Neural-based Program Decompiler

no code implementations28 Jun 2019 Cheng Fu, Huili Chen, Haolan Liu, Xinyun Chen, Yuandong Tian, Farinaz Koushanfar, Jishen Zhao

Reverse engineering of binary executables is a critical problem in the computer security domain.

Computer Security Malware Detection

Towards Safety-Aware Computing System Design in Autonomous Vehicles

no code implementations21 May 2019 Hengyu Zhao, Yubo Zhang, Pingfan Meng, Hui Shi, Li Erran Li, Tiancheng Lou, Jishen Zhao

To address this issue, we propose a `safety score' as a primary metric for measuring the level of safety in AV computing system design.

Autonomous Driving Management

Towards Fast and Energy-Efficient Binarized Neural Network Inference on FPGA

no code implementations4 Oct 2018 Cheng Fu, Shilin Zhu, Hao Su, Ching-En Lee, Jishen Zhao

Thus there does exist redundancy that can be exploited to further reduce the amount of on-chip computations.

Cannot find the paper you are looking for? You can Submit a new open access paper.