Search Results for author: Qingcheng Zeng

Found 17 papers, 8 papers with code

Fancy Man Launches Zippo at WNUT 2020 Shared Task-1: A Bert Case Model for Wet Lab Entity Extraction

no code implementations EMNLP (WNUT) 2020 Qingcheng Zeng, Xiaoyang Fang, Zhexin Liang, Haoding Meng

Automatic or semi-automatic conversion of protocols specifying steps in performing a lab procedure into machine-readable format benefits biological research a lot.

Exploring Multilingual Probing in Large Language Models: A Cross-Language Analysis

no code implementations22 Sep 2024 Daoyang Li, Mingyu Jin, Qingcheng Zeng, Haiyan Zhao, Mengnan Du

Probing techniques for large language models (LLMs) have primarily focused on English, overlooking the vast majority of the world's languages.

Uncertainty is Fragile: Manipulating Uncertainty in Large Language Models

2 code implementations15 Jul 2024 Qingcheng Zeng, Mingyu Jin, Qinkai Yu, Zhenting Wang, Wenyue Hua, ZiHao Zhou, Guangyan Sun, Yanda Meng, Shiqing Ma, Qifan Wang, Felix Juefei-Xu, Kaize Ding, Fan Yang, Ruixiang Tang, Yongfeng Zhang

We demonstrate that an attacker can embed a backdoor in LLMs, which, when activated by a specific trigger in the input, manipulates the model's uncertainty without affecting the final output.

Backdoor Attack Multiple-choice

Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?

1 code implementation10 Apr 2024 Mingyu Jin, Qinkai Yu, Jingyuan Huang, Qingcheng Zeng, Zhenting Wang, Wenyue Hua, Haiyan Zhao, Kai Mei, Yanda Meng, Kaize Ding, Fan Yang, Mengnan Du, Yongfeng Zhang

In this paper, we explore the hypothesis that LLMs process concepts of varying complexities in different layers, introducing the idea of ``Concept Depth'' to suggest that more complex concepts are typically acquired in deeper layers.

KG-Rank: Enhancing Large Language Models for Medical QA with Knowledge Graphs and Ranking Techniques

1 code implementation9 Mar 2024 Rui Yang, Haoran Liu, Edison Marrese-Taylor, Qingcheng Zeng, Yu He Ke, Wanxin Li, Lechao Cheng, Qingyu Chen, James Caverlee, Yutaka Matsuo, Irene Li

In this work, we develop an augmented LLM framework, KG-Rank, which leverages a medical knowledge graph (KG) along with ranking and re-ranking techniques, to improve the factuality of long-form question answering (QA) in the medical domain.

Knowledge Graphs Long Form Question Answering +1

Consistent and Asymptotically Statistically-Efficient Solution to Camera Motion Estimation

1 code implementation2 Mar 2024 Guangyang Zeng, Qingcheng Zeng, Xinghan Li, Biqiang Mu, Jiming Chen, Ling Shi, Junfeng Wu

Given 2D point correspondences between an image pair, inferring the camera motion is a fundamental issue in the computer vision community.

Motion Estimation

Large Language Models on Wikipedia-Style Survey Generation: an Evaluation in NLP Concepts

1 code implementation21 Aug 2023 Fan Gao, Hang Jiang, Rui Yang, Qingcheng Zeng, Jinghui Lu, Moritz Blum, Dairui Liu, Tianwei She, Yuang Jiang, Irene Li

Educational materials such as survey articles in specialized fields like computer science traditionally require tremendous expert inputs and are therefore expensive to create and update.

Hallucination Machine Translation +2

Large Language Models Are Partially Primed in Pronoun Interpretation

1 code implementation26 May 2023 Suet-Ying Lam, Qingcheng Zeng, Kexun Zhang, Chenyu You, Rob Voigt

Recent psycholinguistic studies suggest that humans adapt their referential biases with recent exposure to referential patterns; closely replicating three relevant psycholinguistic experiments from Johnson & Arnold (2022) in an in-context learning (ICL) framework, we found that InstructGPT adapts its pronominal interpretations in response to the frequency of referential patterns in the local discourse, though in a limited fashion: adaptation was only observed relative to syntactic but not semantic biases.

In-Context Learning

GreenPLM: Cross-Lingual Transfer of Monolingual Pre-Trained Language Models at Almost No Cost

1 code implementation13 Nov 2022 Qingcheng Zeng, Lucas Garay, Peilin Zhou, Dading Chong, Yining Hua, Jiageng Wu, Yikang Pan, Han Zhou, Rob Voigt, Jie Yang

Large pre-trained models have revolutionized natural language processing (NLP) research and applications, but high training costs and limited data resources have prevented their benefits from being shared equally amongst speakers of all the world's languages.

Cross-Lingual Transfer

Calibrate and Refine! A Novel and Agile Framework for ASR-error Robust Intent Detection

no code implementations23 May 2022 Peilin Zhou, Dading Chong, Helin Wang, Qingcheng Zeng

The past ten years have witnessed the rapid development of text-based intent detection, whose benchmark performances have already been taken to a remarkable level by deep learning techniques.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Fancy Man Lauches Zippo at WNUT 2020 Shared Task-1: A Bert Case Model for Wet Lab Entity Extraction

no code implementations28 Sep 2020 Haoding Meng, Qingcheng Zeng, Xiaoyang Fang, Zhexin Liang

Automatic or semi-automatic conversion of protocols specifying steps in performing a lab procedure into machine-readable format benefits biological research a lot.

Cannot find the paper you are looking for? You can Submit a new open access paper.