1 code implementation • 29 Jul 2024 • Canyu Chen, Baixiang Huang, Zekun Li, Zhaorun Chen, Shiyang Lai, Xiongxiao Xu, Jia-Chen Gu, Jindong Gu, Huaxiu Yao, Chaowei Xiao, Xifeng Yan, William Yang Wang, Philip Torr, Dawn Song, Kai Shu
Then, we find that editing attacks can inject both types of misinformation into LLMs, and the effectiveness is particularly high for commonsense misinformation injection.
1 code implementation • 6 Jul 2024 • Zekun Li, Xianjun Yang, Kyuri Choi, Wanrong Zhu, Ryan Hsieh, HyeonJung Kim, Jin Hyuk Lim, Sungyoung Ji, Byungju Lee, Xifeng Yan, Linda Ruth Petzold, Stephen D. Wilson, Woosang Lim, William Yang Wang
The rapid advancement of Large Language Models (LLMs) and Large Multimodal Models (LMMs) has heightened the demand for AI-based scientific assistants capable of understanding scientific articles and figures.
no code implementations • 13 Jun 2024 • Xuannan Liu, Zekun Li, Peipei Li, Shuhan Xia, Xing Cui, Linzhi Huang, Huaibo Huang, Weihong Deng, Zhaofeng He
Current multimodal misinformation detection (MMD) methods often assume a single source and type of forgery for each sample, which is insufficient for real-world scenarios where multiple forgery sources coexist.
no code implementations • 13 Jun 2024 • Fei Wang, Xingyu Fu, James Y. Huang, Zekun Li, Qin Liu, Xiaogeng Liu, Mingyu Derek Ma, Nan Xu, Wenxuan Zhou, Kai Zhang, Tianyi Lorena Yan, Wenjie Jacky Mo, Hsiang-Hui Liu, Pan Lu, Chunyuan Li, Chaowei Xiao, Kai-Wei Chang, Dan Roth, Sheng Zhang, Hoifung Poon, Muhao Chen
We introduce MuirBench, a comprehensive benchmark that focuses on robust multi-image understanding capabilities of multimodal LLMs.
no code implementations • 1 Jun 2024 • Xing Cui, Peipei Li, Zekun Li, Xuannan Liu, Yueying Zou, Zhaofeng He
Specifically, semantic guidance is derived by establishing a semantic editing direction based on reasoned intentions, while quality guidance is achieved through classifier guidance using an image fidelity discriminator.
no code implementations • 4 Mar 2024 • Xuannan Liu, Peipei Li, Huaibo Huang, Zekun Li, Xing Cui, Jiahao Liang, Lixiong Qin, Weihong Deng, Zhaofeng He
The massive generation of multimodal fake news involving both text and images exhibits substantial distribution discrepancies, prompting the need for generalized detectors.
1 code implementation • 16 Feb 2024 • Zekun Li, Zhiyu Zoey Chen, Mike Ross, Patrick Huber, Seungwhan Moon, Zhaojiang Lin, Xin Luna Dong, Adithya Sagar, Xifeng Yan, Paul A. Crook
We also show that by fine-tuning on a small collection of diverse task-oriented dialogues, we can equip modestly sized models, specifically a 13B parameter LLaMA2-Chat model, with function-calling capabilities and DST performance comparable to ChatGPT while maintaining their chat capabilities.
no code implementations • CVPR 2024 • Chandradeep Pokhariya, Ishaan N Shah, Angela Xing, Zekun Li, Kefan Chen, Avinash Sharma, Srinath Sridhar
Since our representation uses Gaussian primitives, it enables us to efficiently and accurately estimate contacts between the hand and the object.
no code implementations • 28 Nov 2023 • Zhengming Yu, Zhiyang Dou, Xiaoxiao Long, Cheng Lin, Zekun Li, YuAn Liu, Norman Müller, Taku Komura, Marc Habermann, Christian Theobalt, Xin Li, Wenping Wang
The experiments demonstrate the superior performance of Surf-D in shape generation across multiple modalities as conditions.
1 code implementation • 25 Nov 2023 • Xing Cui, Zekun Li, Pei Pei Li, Huaibo Huang, Xuannan Liu, Zhaofeng He
We employ DDIM inversion to extract this noise from the reference image and leverage a diffusion model to generate new stylized images from the "style" noise.
1 code implementation • 23 Oct 2023 • Zekun Li, Wenxuan Zhou, Yao-Yi Chiang, Muhao Chen
This paper introduces GeoLM, a geospatially grounded language model that enhances the understanding of geo-entities in natural language.
1 code implementation • 23 Oct 2023 • Xinlu Zhang, Chenxin Tian, Xianjun Yang, Lichang Chen, Zekun Li, Linda Ruth Petzold
Instruction-finetuning (IFT) has become crucial in aligning Large Language Models (LLMs) with diverse human needs and has shown great potential in medical applications.
no code implementations • 16 Sep 2023 • Zhiyuan Hu, Yue Feng, Yang Deng, Zekun Li, See-Kiong Ng, Anh Tuan Luu, Bryan Hooi
Recently, the development of large language models (LLMs) has been significantly enhanced the question answering and dialogue generation, and makes them become increasingly popular in current practical scenarios.
1 code implementation • ICCV 2023 • Zekun Li, Lei Qi, Yinghuan Shi, Yang Gao
Semi-supervised learning (SSL) aims to leverage massive unlabeled data when labels are expensive to obtain.
2 code implementations • 17 Aug 2023 • Zekun Li, Baolin Peng, Pengcheng He, Xifeng Yan
In this work, we establish a benchmark to evaluate the robustness of instruction-following LLMs against prompt injection attacks.
no code implementations • 29 Jun 2023 • Jina Kim, Zekun Li, Yijun Lin, Min Namgung, Leeje Jang, Yao-Yi Chiang
mapKurator empowers automated extraction, post-processing, and linkage of text labels from large numbers of large-dimension historical map scans.
no code implementations • CVPR 2023 • Fang Zhao, Zekun Li, Shaoli Huang, Junwu Weng, Tianfei Zhou, Guo-Sen Xie, Jue Wang, Ying Shan
Once the anchor transformations are found, per-vertex nonlinear displacements of the garment template can be regressed in a canonical space, which reduces the complexity of deformation space learning.
no code implementations • 20 Mar 2023 • Xing Cui, Zekun Li, Peipei Li, Yibo Hu, Hailin Shi, Zhaofeng He
This paper explores interactive facial image editing via dialogue and introduces the ChatEdit benchmark dataset for evaluating image editing and conversation abilities in this context.
1 code implementation • NeurIPS 2023 • Zekun Li, Shiyang Li, Xifeng Yan
This paper introduces a novel perspective by converting irregularly sampled time series into line graph images, then utilizing powerful pre-trained vision transformers for time series classification in the same way as image classification.
1 code implementation • NeurIPS 2023 • Zekun Li, Baolin Peng, Pengcheng He, Michel Galley, Jianfeng Gao, Xifeng Yan
Our experiments demonstrate that the framework consistently improves LLMs' (e. g., ChatGPT, Codex, InstructGPT) performance on these supervised tasks using minimal labeled data.
no code implementations • ICCV 2023 • Zongyang Ma, Ziqi Zhang, Yuxin Chen, Zhongang Qi, Yingmin Luo, Zekun Li, Chunfeng Yuan, Bing Li, XiaoHu Qie, Ying Shan, Weiming Hu
This paper proposes a novel generative model, Order-Prompted Tag Sequence Generation (OP-TSG), according to the above characteristics.
no code implementations • 21 Oct 2022 • Zekun Li, Jina Kim, Yao-Yi Chiang, Muhao Chen
Characterizing geo-entities is integral to various application domains, such as geo-intelligence and map comprehension, while a key challenge is to capture the spatial-varying context of an entity.
no code implementations • 13 Oct 2022 • Shiyang Li, Jianshu Chen, Yelong Shen, Zhiyu Chen, Xinlu Zhang, Zekun Li, Hong Wang, Jing Qian, Baolin Peng, Yi Mao, Wenhu Chen, Xifeng Yan
Integrating free-text explanations to in-context learning of large language models (LLM) is shown to elicit strong reasoning capabilities along with reasonable explanations.
1 code implementation • 9 Oct 2022 • Zekun Li, Wenhu Chen, Shiyang Li, Hong Wang, Jing Qian, Xifeng Yan
Experimental results on the MultiWOZ dataset demonstrate that training a model on the simulated dialogues leads to even better performance than using the same amount of human-generated dialogues under the challenging low-resource settings, with as few as 85 dialogues as a seed.
no code implementations • 9 Aug 2022 • Jing Qian, Hong Wang, Zekun Li, Shiyang Li, Xifeng Yan
LMs with tutor is able to deliver 100% accuracy in situations of OOD and repeating symbols, shedding new insights on the boundary of large LMs in induction.
no code implementations • 23 Jul 2022 • Ji Liu, Dong Li, Zekun Li, Han Liu, Wenjing Ke, Lu Tian, Yi Shan
Sample assignment plays a prominent part in modern object detection approaches.
1 code implementation • 13 Jul 2022 • Zekun Li, Zhengyang Geng, Zhao Kang, Wenyu Chen, Yibo Yang
To understand the instability in training, we detect the gradient flow of attention and observe gradient conflict among attention branches.
no code implementations • 12 Dec 2021 • Zekun Li, Runyu Guan, Qianmu Yu, Yao-Yi Chiang, Craig A. Knoblock
We show that the state-of-the-art text detection models (e. g., PSENet) can benefit from the synthetic historical maps and achieve significant improvement for historical map text detection.
1 code implementation • 3 Dec 2021 • Zekun Li, Yao-Yi Chiang, Sasan Tavakkol, Basel Shbita, Johannes H. Uhl, Stefan Leyk, Craig A. Knoblock
This paper presents an end-to-end approach to address the real-world problem of finding and indexing historical map images.
Optical Character Recognition Optical Character Recognition (OCR)
no code implementations • 18 Sep 2021 • Zekun Li, Yufan Liu, Bing Li, Weiming Hu, Kebin Wu, Pei Wang
CDI builds the global attention and interaction among different levels in decoupled space which also solves the problem of heavy computation.
1 code implementation • 25 May 2021 • Shu Wu, Zekun Li, Yunyue Su, Zeyu Cui, XiaoYu Zhang, Liang Wang
To solve the problems, we propose a novel approach, Graph Factorization Machine (GraphFM), by naturally representing features in the graph structure.
no code implementations • 7 Apr 2021 • Zeyu Cui, Zekun Li, Shu Wu, XiaoYu Zhang, Qiang Liu, Liang Wang, Mengmeng Ai
We naturally generalizes the embedding propagation scheme of GCN to dynamic setting in an efficient manner, which is to propagate the change along the graph to update node embeddings.
no code implementations • 7 Feb 2021 • Zekun Li, Wei Zhao, Feng Shi, Lei Qi, Xingzhi Xie, Ying WEI, Zhongxiang Ding, Yang Gao, Shangjie Wu, Jun Liu, Yinghuan Shi, Dinggang Shen
How to fast and accurately assess the severity level of COVID-19 is an essential problem, when millions of people are suffering from the pandemic around the world.
1 code implementation • 12 Dec 2020 • Yuliang Guo, Zhong Li, Zekun Li, Xiangyu Du, Shuxue Quan, Yi Xu
In this paper, a real-time method called PoP-Net is proposed to predict multi-person 3D poses from a depth image.
no code implementations • 10 Dec 2020 • Yujia Zheng, Siyi Liu, Zekun Li, Shu Wu
As there is generally no side information in the setting of sequential recommendation task, previous cold-start methods could not be applied when only user-item interactions are available.
no code implementations • 16 Nov 2020 • Zekun Li, Yufan Liu, Bing Li, Weiming Hu
Furthermore, these two components are both plug-and-play and can be embedded in any backbone.
no code implementations • 13 Nov 2020 • Zekun Li, Yujia Zheng, Shu Wu, XiaoYu Zhang, Liang Wang
In this work, we propose to model user-item interactions as a heterogeneous graph which consists of not only user-item edges indicating their interaction but also user-user edges indicating their similarity.
1 code implementation • 21 Sep 2020 • Yujia Zheng, Siyi Liu, Zekun Li, Shu Wu
These item transitions include potential collaborative information and reflect similar behavior patterns, which we assume may help with the recommendation for the target session.
Ranked #6 on Session-Based Recommendations on Diginetica
5 code implementations • 12 Oct 2019 • Zekun Li, Zeyu Cui, Shu Wu, Xiao-Yu Zhang, Liang Wang
The key of this task is to model feature interactions among different feature fields.
Ranked #10 on Click-Through Rate Prediction on Avazu
1 code implementation • 31 Jul 2019 • Zekun Li, Zeyu Cui, Shu Wu, Xiao-Yu Zhang, Liang Wang
To achieve the alignment, we minimize the distances between distributions with unsupervised adversarial learning, and also the distances between some annotated compatible items which play the role of anchor points to help align.
1 code implementation • 21 Feb 2019 • Zeyu Cui, Zekun Li, Shu Wu, Xiao-Yu Zhang, Liang Wang
In this paper, we aim to investigate a practical problem of fashion recommendation by answering the question "which item should we select to match with the given fashion items and form a compatible outfit".
Ranked #1 on Recommendation Systems on Polyvore (Accuracy metric)