Search Results for author: Tong Niu

Found 24 papers, 8 papers with code

JudgeRank: Leveraging Large Language Models for Reasoning-Intensive Reranking

no code implementations31 Oct 2024 Tong Niu, Shafiq Joty, Ye Liu, Caiming Xiong, Yingbo Zhou, Semih Yavuz

Accurate document retrieval is crucial for the success of retrieval-augmented generation (RAG) applications, including open-domain question answering and code completion.

Code Completion Open-Domain Question Answering +3

Improving LLM Reasoning through Scaling Inference Computation with Collaborative Verification

no code implementations5 Oct 2024 Zhenwen Liang, Ye Liu, Tong Niu, Xiangliang Zhang, Yingbo Zhou, Semih Yavuz

Moreover, to leverage the unique strengths of different reasoning strategies, we propose a novel collaborative method integrating Chain-of-Thought (CoT) and Program-of-Thought (PoT) solutions for verification.

GSM8K Math

RAM2C: A Liberal Arts Educational Chatbot based on Retrieval-augmented Multi-role Multi-expert Collaboration

1 code implementation23 Sep 2024 Haoyu Huang, Tong Niu, Rui Yang, Luping Shi

Then, RAM2C organizes LLMs, which are retrieval-augmented by the above different knowledge bases, into multi-experts groups with distinct roles to generate the HTS-compliant educational dialogues dataset.

Chatbot Ethics +1

Mixture of Prompt Learning for Vision Language Models

no code implementations18 Sep 2024 Yu Du, Tong Niu, Rong Zhao

This supervision ensures that the text features derived from soft prompts remain close to those from their corresponding hard prompts, preserving initial knowledge and mitigating overfitting.

Domain Generalization Few-Shot Learning

Solution-oriented Agent-based Models Generation with Verifier-assisted Iterative In-context Learning

no code implementations4 Feb 2024 Tong Niu, Weihao Zhang, Rong Zhao

In SAGE, we introduce an semi-structured conceptual representation expliciting the intricate structures of ABMs and an objective representation to guide LLMs in modeling scenarios and proposing hypothetical solutions through in-context learning.

In-Context Learning

General Automatic Solution Generation of Social Problems

no code implementations25 Jan 2024 Tong Niu, Haoyu Huang, Yu Du, Weihao Zhang, Luping Shi, Rong Zhao

Given the escalating intricacy and multifaceted nature of contemporary social systems, manually generating solutions to address pertinent social issues has become a formidable task.

DIVKNOWQA: Assessing the Reasoning Ability of LLMs via Open-Domain Question Answering over Knowledge Base and Text

no code implementations31 Oct 2023 Wenting Zhao, Ye Liu, Tong Niu, Yao Wan, Philip S. Yu, Shafiq Joty, Yingbo Zhou, Semih Yavuz

Moreover, a significant gap in the current landscape is the absence of a realistic benchmark for evaluating the effectiveness of grounding LLMs on heterogeneous knowledge sources (e. g., knowledge base and text).

Knowledge Graphs Open-Domain Question Answering +2

XGen-7B Technical Report

1 code implementation7 Sep 2023 Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause, Senthil Purushwalkam, Tong Niu, Wojciech Kryściński, Lidiya Murakhovs'ka, Prafulla Kumar Choubey, Alex Fabbri, Ye Liu, Rui Meng, Lifu Tu, Meghana Bhat, Chien-Sheng Wu, Silvio Savarese, Yingbo Zhou, Shafiq Joty, Caiming Xiong

Most open-source LLMs, on the other hand, are limited in their ability to support longer sequence lengths, which is a key requirement for many tasks that require inference over an input context.

2k 8k

Attention-based 3D CNN with Multi-layer Features for Alzheimer's Disease Diagnosis using Brain Images

no code implementations10 Aug 2023 Yanteng Zhang, Qizhi Teng, Xiaohai He, Tong Niu, Lipei Zhang, Yan Liu, Chao Ren

Structural MRI and PET imaging play an important role in the diagnosis of Alzheimer's disease (AD), showing the morphological changes and glucose metabolism changes in the brain respectively.

OneAligner: Zero-shot Cross-lingual Transfer with One Rich-Resource Language Pair for Low-Resource Sentence Retrieval

no code implementations Findings (ACL) 2022 Tong Niu, Kazuma Hashimoto, Yingbo Zhou, Caiming Xiong

When finetuned on a single rich-resource language pair, be it English-centered or not, our model is able to match the performance of the ones finetuned on all language pairs under the same data budget with less than 2. 0 points decrease in accuracy.

Machine Translation Retrieval +3

Char2Subword: Extending the Subword Embedding Space Using Robust Character Compositionality

no code implementations Findings (EMNLP) 2021 Gustavo Aguilar, Bryan McCann, Tong Niu, Nazneen Rajani, Nitish Keskar, Thamar Solorio

To alleviate these challenges, we propose a character-based subword module (char2subword) that learns the subword embedding table in pre-trained models like BERT.

CoCo: Controllable Counterfactuals for Evaluating Dialogue State Trackers

2 code implementations ICLR 2021 Shiyang Li, Semih Yavuz, Kazuma Hashimoto, Jia Li, Tong Niu, Nazneen Rajani, Xifeng Yan, Yingbo Zhou, Caiming Xiong

Dialogue state trackers have made significant progress on benchmark datasets, but their generalization capability to novel and realistic scenarios beyond the held-out conversations is less understood.

Ranked #2 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.1 (using extra training data)

counterfactual Dialogue State Tracking +1

Unsupervised Paraphrasing with Pretrained Language Models

no code implementations EMNLP 2021 Tong Niu, Semih Yavuz, Yingbo Zhou, Nitish Shirish Keskar, Huan Wang, Caiming Xiong

To enforce a surface form dissimilar from the input, whenever the language model emits a token contained in the source sequence, DB prevents the model from outputting the subsequent source token for the next generation step.

Blocking Language Modelling +3

Explaining and Improving Model Behavior with k Nearest Neighbor Representations

no code implementations18 Oct 2020 Nazneen Fatema Rajani, Ben Krause, Wengpeng Yin, Tong Niu, Richard Socher, Caiming Xiong

Interpretability techniques in NLP have mainly focused on understanding individual predictions using attention visualization or gradient-based saliency maps over tokens.

Natural Language Inference

AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses

no code implementations15 Jan 2020 Tong Niu, Mohit Bansal

In our work, we build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.

Diversity Feature Engineering +1

Automatically Learning Data Augmentation Policies for Dialogue Tasks

1 code implementation IJCNLP 2019 Tong Niu, Mohit Bansal

Automatic data augmentation (AutoAugment) (Cubuk et al., 2019) searches for optimal perturbation policies via a controller trained using performance rewards of a sampled policy on the target task, hence reducing data-level model bias.

Data Augmentation Dialogue Generation +2

Deleter: Leveraging BERT to Perform Unsupervised Successive Text Compression

no code implementations7 Sep 2019 Tong Niu, Caiming Xiong, Richard Socher

In this work, we propose a fully unsupervised model, Deleter, that is able to discover an "optimal deletion path" for an arbitrary sentence, where each intermediate sequence along the path is a coherent subsequence of the previous one.

Language Modelling Reading Comprehension +3

A novel hybrid model based on multi-objective Harris hawks optimization algorithm for daily PM2.5 and PM10 forecasting

no code implementations30 May 2019 Pei Du, Jianzhou Wang, Yan Hao, Tong Niu, Wendong Yang

Next, a new multi-objective algorithm called MOHHO is first developed in this study, which are introduced to tune the parameters of ELM model with high forecasting accuracy and stability for air pollution series prediction, simultaneously.

Air Pollution Prediction Time Series +1

Adversarial Over-Sensitivity and Over-Stability Strategies for Dialogue Models

1 code implementation CONLL 2018 Tong Niu, Mohit Bansal

We present two categories of model-agnostic adversarial strategies that reveal the weaknesses of several generative, task-oriented dialogue models: Should-Not-Change strategies that evaluate over-sensitivity to small and semantics-preserving edits, as well as Should-Change strategies that test if a model is over-stable against subtle yet semantics-changing modifications.

Polite Dialogue Generation Without Parallel Data

1 code implementation TACL 2018 Tong Niu, Mohit Bansal

We present three weakly-supervised models that can generate diverse polite (or rude) dialogue responses without parallel data.

Decoder Dialogue Generation +4

Detecting Linguistic Characteristics of Alzheimer's Dementia by Interpreting Neural Models

no code implementations NAACL 2018 Sweta Karlekar, Tong Niu, Mohit Bansal

More importantly, we next interpret what these neural models have learned about the linguistic characteristics of AD patients, via analysis based on activation clustering and first-derivative saliency techniques.

Clustering

Cannot find the paper you are looking for? You can Submit a new open access paper.