Search Results for author: Leon Bergen

Found 15 papers, 8 papers with code

Predicting Reference: What do Language Models Learn about Discourse Models?

no code implementations EMNLP 2020 Shiva Upadhye, Leon Bergen, Andrew Kehler

Whereas there is a growing literature that probes neural language models to assess the degree to which they have latently acquired grammatical knowledge, little if any research has investigated their acquisition of discourse modeling ability.

Measuring Risk of Bias in Biomedical Reports: The RoBBR Benchmark

1 code implementation28 Nov 2024 Jianyou Wang, Weili Cao, Longtian Bao, Youze Zheng, Gil Pasternak, Kaicheng Wang, Xiaoyue Wang, Ramamohan Paturi, Leon Bergen

To draw reliable conclusions, these systems should take into account the quality of available evidence, placing more weight on studies that use a valid methodology.

valid

Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation

no code implementations1 Nov 2024 Bohan Lyu, Yadi Cao, Duncan Watson-Parris, Leon Bergen, Taylor Berg-Kirkpatrick, Rose Yu

On average, our models demonstrate a 28. 18% improvement in answer accuracy and a 13. 89% increase in tool usage precision across all datasets, surpassing state-of-the-art models including GPT-4o and Claude-3. 5.

Epidemiology Knowledge Distillation +1

ClimaQA: An Automated Evaluation Framework for Climate Foundation Models

no code implementations22 Oct 2024 Veeramakali Vignesh Manivannan, Yasaman Jafari, Srikar Eranky, Spencer Ho, Rose Yu, Duncan Watson-Parris, Yian Ma, Leon Bergen, Taylor Berg-Kirkpatrick

However, a critical issue remains: the lack of a comprehensive evaluation framework capable of assessing the quality and scientific validity of model outputs.

Dissociation of Faithful and Unfaithful Reasoning in LLMs

1 code implementation23 May 2024 Evelyn Yee, Alice Li, Chenyu Tang, Yeon Ho Jung, Ramamohan Paturi, Leon Bergen

Through analysis of error recovery behaviors, we find evidence for unfaithfulness in Chain of Thought, which occurs when models arrive at the correct answer despite invalid reasoning text.

IR2: Information Regularization for Information Retrieval

1 code implementation25 Feb 2024 Jianyou Wang, Kaicheng Wang, Xiaoyue Wang, Weili Cao, Ramamohan Paturi, Leon Bergen

This approach, representing a novel application of regularization techniques in synthetic data creation for IR, is tested on three recent IR tasks characterized by complex queries: DORIS-MAE, ArguAna, and WhatsThatBook.

Information Retrieval Retrieval +1

DORIS-MAE: Scientific Document Retrieval using Multi-level Aspect-based Queries

1 code implementation7 Oct 2023 Jianyou Wang, Kaicheng Wang, Xiaoyue Wang, Prudhviraj Naidu, Leon Bergen, Ramamohan Paturi

In scientific research, the ability to effectively retrieve relevant documents based on complex, multifaceted queries is critical.

Retrieval

Systematic Generalization with Edge Transformers

1 code implementation NeurIPS 2021 Leon Bergen, Timothy J. O'Donnell, Dzmitry Bahdanau

Recent research suggests that systematic generalization in natural language understanding remains a challenge for state-of-the-art neural models such as Transformers and Graph Neural Networks.

Dependency Parsing Natural Language Understanding +3

Jointly Learning Truth-Conditional Denotations and Groundings using Parallel Attention

no code implementations14 Apr 2021 Leon Bergen, Dzmitry Bahdanau, Timothy J. O'Donnell

We present a model that jointly learns the denotations of words together with their groundings using a truth-conditional semantics.

Question Answering Visual Question Answering

Word Frequency Does Not Predict Grammatical Knowledge in Language Models

1 code implementation EMNLP 2020 Charles Yu, Ryan Sie, Nico Tedeschi, Leon Bergen

Neural language models learn, to varying degrees of accuracy, the grammatical properties of natural languages.

Speakers enhance contextually confusable words

no code implementations ACL 2020 Eric Meinhardt, Eric Bakovic, Leon Bergen

Recent work has found evidence that natural languages are shaped by pressures for efficient communication {---} e. g. the more contextually predictable a word is, the fewer speech sounds or syllables it has (Piantadosi et al. 2011).

Cannot find the paper you are looking for? You can Submit a new open access paper.