no code implementations • 21 Jan 2024 • Rongqing Cong, Wenyang He, Mingxuan Li, Bangning Luo, Zebin Yang, Yuchao Yang, Ru Huang, Bonan Yan
Large language models (LLMs) with Transformer architectures have become phenomenal in natural language processing, multimodal generative artificial intelligence, and agent-oriented artificial intelligence.
1 code implementation • NeurIPS 2023 • Jungtaek Kim, Mingxuan Li, Oliver Hinder, Paul W. Leu
To design and understand these nanophotonic structures, electrodynamic simulations are essential.
no code implementations • 14 Apr 2023 • Jingyuan Wang, Yufan Wu, Mingxuan Li, Xin Lin, Junjie Wu, Chao Li
While having achieved great success in rich real-life applications, deep neural network (DNN) models have long been criticized for their vulnerability to adversarial attacks.
no code implementations • 31 Aug 2022 • Bo Ye, Feng Li, Mingxuan Li, Peipei Yan, Huiting Yang, Lihua Wang
Based on the end-to-end learning paradigm, this paper proposes an intelligent detection method for substation insulator defects based on CenterMask.
no code implementations • 9 Dec 2021 • Yiheng Xie, Mingxuan Li, Shangqun Yu, Michael Littman
Though deep reinforcement learning agents have achieved unprecedented success in recent years, their learned policies can be brittle, failing to generalize to even slight modifications of their environments or unfamiliar situations.
no code implementations • 18 Jan 2021 • Mingxuan Li, Yuanxun Shao
Artificial neural network has achieved the state-of-art performance in fault detection on the Tennessee Eastman process, but it often requires enormous memory to fund its massive parameters.
1 code implementation • 7 Aug 2020 • Mingxuan Li, Michael L. Littman
We demonstrate the potential of graph neural network in supporting sample efficient learning by showing that Deep Graph Value Network can outperform unstructured baselines by a large margin in solving the Markov Decision Process (MDP).
no code implementations • 28 Feb 2020 • Mingxuan Li, Jingyuan Wang, Yufan Wu
That is, adversarial examples generated by $L_2$ attacks usually have larger input sensitivity which can be used to identify them efficiently.