Search Results for author: Tao Ji

Found 22 papers, 13 papers with code

AntLM: Bridging Causal and Masked Language Models

no code implementations4 Dec 2024 Xinru Yu, Bin Guo, Shiwei Luo, Jie Wang, Tao Ji, Yuanbin Wu

For the BabyLM Challenge 2024, we propose a novel language modeling paradigm named $\textbf{AntLM}$, which integrates both CLM and MLM to leverage the advantages of these two classic paradigms.

Causal Language Modeling Decoder +2

Multi-Programming Language Sandbox for LLMs

1 code implementation30 Oct 2024 Shihan Dou, Jiazheng Zhang, Jianxiang Zang, Yunbo Tao, Weikang Zhou, Haoxiang Jia, Shichun Liu, Yuming Yang, Zhiheng Xi, Shenxi Wu, Shaoqing Zhang, Muling Wu, Changze Lv, Limao Xiong, WenYu Zhan, Lin Zhang, Rongxiang Weng, Jingang Wang, Xunliang Cai, Yueming Wu, Ming Wen, Rui Zheng, Tao Ji, Yixin Cao, Tao Gui, Xipeng Qiu, Qi Zhang, Xuanjing Huang

We introduce MPLSandbox, an out-of-the-box multi-programming language sandbox designed to provide unified and comprehensive feedback from compiler and analysis tools for Large Language Models (LLMs).

Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs

no code implementations15 Oct 2024 Shuo Li, Tao Ji, Xiaoran Fan, Linsheng Lu, Leyi Yang, Yuming Yang, Zhiheng Xi, Rui Zheng, Yuran Wang, Xiaohui Zhao, Tao Gui, Qi Zhang, Xuanjing Huang

Our findings indicate that the ability to prevent sycophancy is predominantly observed in higher layers of the model.

Hallucination

Generation with Dynamic Vocabulary

1 code implementation11 Oct 2024 Yanting Liu, Tao Ji, Changzhi Sun, Yuanbin Wu, Xiaoling Wang

The dynamic vocabulary can be deployed in a plug-and-play way, thus is attractive for various downstream applications.

Language Modelling Question Answering

Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) Models

1 code implementation4 Oct 2024 Yufang Liu, Tao Ji, Changzhi Sun, Yuanbin Wu, Aimin Zhou

Large Vision-Language Models (LVLMs) have achieved impressive performance, yet research has pointed out a serious issue with object hallucinations within these models.

counterfactual Data Augmentation +3

LongHeads: Multi-Head Attention is Secretly a Long Context Processor

1 code implementation16 Feb 2024 Yi Lu, Xin Zhou, wei he, Jun Zhao, Tao Ji, Tao Gui, Qi Zhang, Xuanjing Huang

Instead of allowing each head to attend to the full sentence, which struggles with generalizing to longer sequences due to out-of-distribution (OOD) issues, we allow each head to process in-distribution length by selecting and attending to important context chunks.

Sentence

ToolEyes: Fine-Grained Evaluation for Tool Learning Capabilities of Large Language Models in Real-world Scenarios

1 code implementation1 Jan 2024 Junjie Ye, Guanyu Li, Songyang Gao, Caishuang Huang, Yilong Wu, Sixian Li, Xiaoran Fan, Shihan Dou, Tao Ji, Qi Zhang, Tao Gui, Xuanjing Huang

Existing evaluations of tool learning primarily focus on validating the alignment of selected tools for large language models (LLMs) with expected outcomes.

FOSS: A Self-Learned Doctor for Query Optimizer

no code implementations11 Dec 2023 Kai Zhong, Luming Sun, Tao Ji, Cuiping Li, Hong Chen

They either learn to construct plans from scratch in a bottom-up manner or steer the plan generation behavior of traditional optimizer using hints.

Deep Reinforcement Learning

Evaluator for Emotionally Consistent Chatbots

1 code implementation2 Dec 2021 Chenxiao Liu, Guanzhi Deng, Tao Ji, Difei Tang, Silai Zheng

One challenge for evaluating current sequence- or dialogue-level chatbots, such as Empathetic Open-domain Conversation Models, is to determine whether the chatbot performs in an emotionally consistent way.

Chatbot Diversity

Generating CCG Categories

1 code implementation15 Mar 2021 Yufang Liu, Tao Ji, Yuanbin Wu, Man Lan

Previous CCG supertaggers usually predict categories using multi-class classification.

Multi-class Classification Sentence

Graph-based Dependency Parsing with Graph Neural Networks

1 code implementation ACL 2019 Tao Ji, Yuanbin Wu, Man Lan

We investigate the problem of efficiently incorporating high-order features into neural graph-based dependency parsing.

Dependency Parsing

A Fast and Lightweight System for Multilingual Dependency Parsing

no code implementations CONLL 2017 Tao Ji, Yuanbin Wu, Man Lan

We present a multilingual dependency parser with a bidirectional-LSTM (BiLSTM) feature extractor and a multi-layer perceptron (MLP) classifier.

Dependency Parsing

Cannot find the paper you are looking for? You can Submit a new open access paper.