no code implementations • 3 Nov 2024 • Yukun Li, Sijia Wang, Lifu Huang, Li-Ping Liu
We propose a novel method combining the LLM's self-consistency with labeled data and training an auxiliary model to estimate the correctness of its responses to questions.
no code implementations • 29 Oct 2024 • Renze Lou, Hanzi Xu, Sijia Wang, Jiangshu Du, Ryo Kamoi, Xiaoxin Lu, Jian Xie, Yuxuan Sun, Yusen Zhang, Jihyun Janice Ahn, Hongchao Fang, Zhuoyang Zou, Wenchao Ma, Xi Li, Kai Zhang, Congying Xia, Lifu Huang, Wenpeng Yin
Numerous studies have assessed the proficiency of AI systems, particularly large language models (LLMs), in facilitating everyday tasks such as email writing, question answering, and creative content generation.
no code implementations • 26 Oct 2024 • Mohammad Beigi, Sijia Wang, Ying Shen, Zihao Lin, Adithya Kulkarni, Jianfeng He, Feng Chen, Ming Jin, Jin-Hee Cho, Dawei Zhou, Chang-Tien Lu, Lifu Huang
In recent years, Large Language Models (LLMs) have become fundamental to a broad spectrum of artificial intelligence applications.
2 code implementations • 21 Sep 2024 • Sijia Wang, Chen Wang, Zhenhao Zhao, Jiqiang Zhang, Weiran Cai
In this work, we first propose a method for measuring exposure bias in SGMs used for molecular conformation generation, which confirms the significant existence of exposure bias in these models and measures its value.
1 code implementation • 12 Sep 2024 • Zhenhao Zhao, Minhong Zhu, Chen Wang, Sijia Wang, Jiqiang Zhang, Li Chen, Weiran Cai
Graph Contrastive Learning (GCL) seeks to learn nodal or graph representations that contain maximal consistent information from graph-structured data.
no code implementations • 19 Jul 2024 • Hanwen Zheng, Sijia Wang, Chris Thomas, Lifu Huang
Chart comprehension presents significant challenges for machine learning models due to the diverse and intricate shapes of charts.
no code implementations • 18 Jun 2024 • Sijia Wang, Lifu Huang
We propose a multi-agent debate as optimization (DAO) system for event extraction, where the primary objective is to iteratively refine the large language models (LLMs) outputs through debating without parameter tuning.
no code implementations • 14 May 2024 • Sijia Wang, Lifu Huang
Addressing the challenge of low-resource information extraction remains an ongoing issue due to the inherent information scarcity within limited training examples.
no code implementations • 12 Apr 2024 • Yingchaojie Feng, Zhizhang Chen, Zhining Kang, Sijia Wang, Minfeng Zhu, Wei zhang, Wei Chen
Addressing these concerns necessitates a comprehensive analysis of jailbreak prompts to evaluate LLMs' defensive capabilities and identify potential weaknesses.
no code implementations • 23 Sep 2023 • Hanwen Zheng, Sijia Wang, Lifu Huang
Document-level information extraction (IE) is a crucial task in natural language processing (NLP).
1 code implementation • 18 Jul 2023 • Yingchaojie Feng, Xingbo Wang, Kam Kwai Wong, Sijia Wang, Yuhong Lu, Minfeng Zhu, Baicheng Wang, Wei Chen
Generative text-to-image models have gained great popularity among the public for their powerful capability to generate high-quality images based on natural language prompts.
no code implementations • 27 May 2023 • Sijia Wang, Alexander Hanbo Li, Henry Zhu, Sheng Zhang, Chung-Wei Hang, Pramuditha Perera, Jie Ma, William Wang, Zhiguo Wang, Vittorio Castelli, Bing Xiang, Patrick Ng
Entities can be expressed in diverse formats, such as texts, images, or column names and cell values in tables.
1 code implementation • 24 May 2023 • Pritika Ramu, Sijia Wang, Lalla Mouatadid, Joy Rimchala, Lifu Huang
Current research in form understanding predominantly relies on large pre-trained language models, necessitating extensive data for pre-training.
no code implementations • 24 May 2023 • Barry Menglong Yao, Yu Chen, Qifan Wang, Sijia Wang, Minqian Liu, Zhiyang Xu, Licheng Yu, Lifu Huang
We propose attribute-aware multimodal entity linking, where the input is a mention described with a text and image, and the goal is to predict the corresponding target entity from a multimodal knowledge base (KB) where each entity is also described with a text description, a visual image and a set of attributes and values.
no code implementations • 13 Jan 2023 • Kaiwen Wan, Lei LI, Dengqiang Jia, Shangqi Gao, Wei Qian, Yingzhi Wu, Huandong Lin, Xiongzheng Mu, Xin Gao, Sijia Wang, Fuping Wu, Xiahai Zhuang
This is particularly evident for the learning-based multi-target landmark detection, where algorithms could be misleading to learn primarily the variation of background due to the varying FOV, failing the detection of targets.
no code implementations • 11 Oct 2022 • Sijia Wang, Yoojin Choi, Junya Chen, Mostafa El-Khamy, Ricardo Henao
This results in the eventual prohibitive expansion of the knowledge repository if we consider learning from a long sequence of tasks.
no code implementations • 14 Apr 2022 • Sijia Wang, Mo Yu, Lifu Huang
We compare various forms of prompts to represent event types and develop a unified framework to incorporate the event type specific prompts for supervised, few-shot, and zero-shot event detection.
no code implementations • 4 Nov 2021 • Junya Chen, Sijia Wang, Lawrence Carin, Chenyang Tao
Distributed learning has become an integral tool for scaling up machine learning and addressing the growing need for data privacy.
no code implementations • Findings (ACL) 2022 • Sijia Wang, Mo Yu, Shiyu Chang, Lichao Sun, Lifu Huang
Event extraction is typically modeled as a multi-class classification problem where event types and argument roles are treated as atomic symbols.
no code implementations • 4 Jun 2021 • Chaofan Chen, Kangcheng Lin, Cynthia Rudin, Yaron Shaposhnik, Sijia Wang, Tong Wang
We propose a framework for such decisions, including a globally interpretable machine learning model, an interactive visualization of it, and several types of summaries and explanations for any given decision.
1 code implementation • NeurIPS 2020 • Yulai Cong, Miaoyun Zhao, Jianqiao Li, Sijia Wang, Lawrence Carin
As a fundamental issue in lifelong learning, catastrophic forgetting is directly caused by inaccessible historical data; accordingly, if the data (information) were memorized perfectly, no forgetting should be expected.
no code implementations • 30 Nov 2018 • Chaofan Chen, Kangcheng Lin, Cynthia Rudin, Yaron Shaposhnik, Sijia Wang, Tong Wang
We propose a possible solution to a public challenge posed by the Fair Isaac Corporation (FICO), which is to provide an explainable model for credit risk assessment.