Search Results for author: Peng Shi

Found 35 papers, 16 papers with code

OpenAgents: An Open Platform for Language Agents in the Wild

2 code implementations16 Oct 2023 Tianbao Xie, Fan Zhou, Zhoujun Cheng, Peng Shi, Luoxuan Weng, Yitao Liu, Toh Jing Hua, Junning Zhao, Qian Liu, Che Liu, Leo Z. Liu, Yiheng Xu, Hongjin Su, Dongchan Shin, Caiming Xiong, Tao Yu

Language agents show potential in being capable of utilizing natural language for varied and intricate tasks in diverse environments, particularly when built upon large language models (LLMs).

2D Object Detection

Binding Language Models in Symbolic Languages

1 code implementation6 Oct 2022 Zhoujun Cheng, Tianbao Xie, Peng Shi, Chengzu Li, Rahul Nadkarni, Yushi Hu, Caiming Xiong, Dragomir Radev, Mari Ostendorf, Luke Zettlemoyer, Noah A. Smith, Tao Yu

We propose Binder, a training-free neural-symbolic framework that maps the task input to a program, which (1) allows binding a unified API of language model (LM) functionalities to a programming language (e. g., SQL, Python) to extend its grammar coverage and thus tackle more diverse questions, (2) adopts an LM as both the program parser and the underlying model called by the API during execution, and (3) requires only a few in-context exemplar annotations.

Language Modelling Semantic Parsing +1

Learning Contextual Representations for Semantic Parsing with Generation-Augmented Pre-Training

3 code implementations18 Dec 2020 Peng Shi, Patrick Ng, Zhiguo Wang, Henghui Zhu, Alexander Hanbo Li, Jun Wang, Cicero Nogueira dos santos, Bing Xiang

Most recently, there has been significant interest in learning contextual representations for various NLP tasks, by leveraging large scale text corpora to train large neural language models with self-supervised learning objectives, such as Masked Language Model (MLM).

Ranked #6 on Text-To-SQL on spider (Exact Match Accuracy (Dev) metric)

Language Modelling Self-Supervised Learning +2

Mr. TyDi: A Multi-lingual Benchmark for Dense Retrieval

1 code implementation EMNLP (MRL) 2021 Xinyu Zhang, Xueguang Ma, Peng Shi, Jimmy Lin

We present Mr. TyDi, a multi-lingual benchmark dataset for mono-lingual retrieval in eleven typologically diverse languages, designed to evaluate ranking with learned dense representations.

Representation Learning Retrieval

Did You Ask a Good Question? A Cross-Domain Question Intention Classification Benchmark for Text-to-SQL

1 code implementation23 Oct 2020 Yusen Zhang, Xiangyu Dong, Shuaichen Chang, Tao Yu, Peng Shi, Rui Zhang

Neural models have achieved significant results on the text-to-SQL task, in which most current work assumes all the input questions are legal and generates a SQL query for any input.

Text-To-SQL

Farewell Freebase: Migrating the SimpleQuestions Dataset to DBpedia

1 code implementation COLING 2018 Michael Azmy, Peng Shi, Jimmy Lin, Ihab Ilyas

To address this problem, we present SimpleDBpediaQA, a new benchmark dataset for simple question answering over knowledge graphs that was created by mapping SimpleQuestions entities and predicates from Freebase to DBpedia.

Knowledge Graphs Question Answering +1

Logic-Consistency Text Generation from Semantic Parses

1 code implementation Findings (ACL) 2021 Chang Shu, Yusen Zhang, Xiangyu Dong, Peng Shi, Tao Yu, Rui Zhang

Text generation from semantic parses is to generate textual descriptions for formal representation inputs such as logic forms and SQL queries.

Text Generation

Segatron: Segment-Aware Transformer for Language Modeling and Understanding

1 code implementation30 Apr 2020 He Bai, Peng Shi, Jimmy Lin, Yuqing Xie, Luchen Tan, Kun Xiong, Wen Gao, Ming Li

To verify this, we propose a segment-aware Transformer (Segatron), by replacing the original token position encoding with a combined position encoding of paragraph, sentence, and token.

Language Modelling Masked Language Modeling +3

Better Language Model with Hypernym Class Prediction

1 code implementation ACL 2022 He Bai, Tong Wang, Alessandro Sordoni, Peng Shi

Class-based language models (LMs) have been long devised to address context sparsity in $n$-gram LMs.

Language Modelling

Video-Text Retrieval by Supervised Sparse Multi-Grained Learning

1 code implementation19 Feb 2023 Yimu Wang, Peng Shi

While recent progress in video-text retrieval has been advanced by the exploration of better representation learning, in this paper, we present a novel multi-grained sparse learning framework, S3MA, to learn an aligned sparse space shared between the video and the text for video-text retrieval.

Representation Learning Retrieval +3

Unified Low-Resource Sequence Labeling by Sample-Aware Dynamic Sparse Finetuning

1 code implementation7 Nov 2023 Sarkar Snigdha Sarathi Das, Ranran Haoran Zhang, Peng Shi, Wenpeng Yin, Rui Zhang

Unfortunately, this requires formatting them into specialized augmented format unknown to the base pretrained language model (PLMs) necessitating finetuning to the target format.

In-Context Learning Language Modelling +6

Strong Baselines for Simple Question Answering over Knowledge Graphs with and without Neural Networks

no code implementations NAACL 2018 Salman Mohammed, Peng Shi, Jimmy Lin

We examine the problem of question answering over knowledge graphs, focusing on simple questions that can be answered by the lookup of a single fact.

Entity Linking Knowledge Graphs +1

Simple Attention-Based Representation Learning for Ranking Short Social Media Posts

no code implementations NAACL 2019 Peng Shi, Jinfeng Rao, Jimmy Lin

This paper explores the problem of ranking short social media posts with respect to user queries using neural networks.

Representation Learning

Cross-Lingual Relevance Transfer for Document Retrieval

no code implementations8 Nov 2019 Peng Shi, Jimmy Lin

Recent work has shown the surprising ability of multi-lingual BERT to serve as a zero-shot cross-lingual transfer model for a number of language processing tasks.

Retrieval Sentence +1

Derivation of Elastic Wave Equation from New Motion Description

no code implementations27 May 2020 Peng Shi

In classical mechanics, the motion of an object is described with Newton's three laws of motion, which means that the motion of the material elements composing a continuum can be described with the particle model.

Classical Physics Materials Science Fluid Dynamics

Cross-Lingual Training of Neural Models for Document Ranking

no code implementations Findings of the Association for Computational Linguistics 2020 Peng Shi, He Bai, Jimmy Lin

We tackle the challenge of cross-lingual training of neural document ranking models for mono-lingual retrieval, specifically leveraging relevance judgments in English to improve search in non-English languages.

Document Ranking Retrieval

End-to-End Cross-Domain Text-to-SQL Semantic Parsing with Auxiliary Task

no code implementations17 Jun 2021 Peng Shi, Tao Yu, Patrick Ng, Zhiguo Wang

Furthermore, we propose two value filling methods to build the bridge from the existing zero-shot semantic parsers to real-world applications, considering most of the existing parsers ignore the values filling in the synthesized SQL.

Semantic Parsing Text-To-SQL

Prefix-to-SQL: Text-to-SQL Generation from Incomplete User Questions

no code implementations15 Sep 2021 Naihao Deng, Shuaichen Chang, Peng Shi, Tao Yu, Rui Zhang

Existing text-to-SQL research only considers complete questions as the input, but lay-users might strive to formulate a complete question.

Text-To-SQL

Hierarchical Character Tagger for Short Text Spelling Error Correction

no code implementations WNUT (ACL) 2021 Mengyi Gao, Canran Xu, Peng Shi

State-of-the-art approaches to spelling error correction problem include Transformer-based Seq2Seq models, which require large training sets and suffer from slow inference time; and sequence labeling models based on Transformer encoders like BERT, which involve token-level label space and therefore a large pre-defined vocabulary dictionary.

Language Modelling

Semi-global Periodic Event-triggered Output Regulation for Nonlinear Multi-agent Systems

no code implementations4 Jan 2022 Shiqi Zheng, Peng Shi, Huiyan Zhang

This study focuses on periodic event-triggered (PET) cooperative output regulation problem for a class of nonlinear multi-agent systems.

XRICL: Cross-lingual Retrieval-Augmented In-Context Learning for Cross-lingual Text-to-SQL Semantic Parsing

no code implementations25 Oct 2022 Peng Shi, Rui Zhang, He Bai, Jimmy Lin

We also include global translation exemplars for a target language to facilitate the translation process for large language models.

In-Context Learning Retrieval +4

Gradient-Based Word Substitution for Obstinate Adversarial Examples Generation in Language Models

no code implementations24 Jul 2023 Yimu Wang, Peng Shi, Hongyang Zhang

Furthermore, to show the transferability of obstinate word substitutions found by GradObstinate, we replace the words in four representative NLP benchmarks with their obstinate substitutions.

Memorization MRPC +1

A novel integrated method of detection-grasping for specific object based on the box coordinate matching

no code implementations20 Jul 2023 Zongmin Liu, Jirui Wang, Jie Li, Zufeng Li, Kai Ren, Peng Shi

Furthermore, a detection-grasping integrated algorithm based on box coordinate matching (DG-BCM) is proposed to obtain the fusion model of object detection and grasp estimation.

Instance Segmentation Object +3

Construction of Paired Knowledge Graph-Text Datasets Informed by Cyclic Evaluation

no code implementations20 Sep 2023 Ali Mousavi, Xin Zhan, He Bai, Peng Shi, Theo Rekatsinas, Benjamin Han, Yunyao Li, Jeff Pound, Josh Susskind, Natalie Schluter, Ihab Ilyas, Navdeep Jaitly

Guided by these observations, we construct a new, improved dataset called LAGRANGE using heuristics meant to improve equivalence between KG and text and show the impact of each of the heuristics on cyclic evaluation.

Hallucination Knowledge Graphs

Cannot find the paper you are looking for? You can Submit a new open access paper.