Search Results for author: Joshua Bensemann

Found 7 papers, 2 papers with code

Do Smaller Language Models Answer Contextualised Questions Through Memorisation Or Generalisation?

no code implementations21 Nov 2023 Tim Hartill, Joshua Bensemann, Michael Witbrock, Patricia J. Riddle

We train two Language Models in a multitask fashion whereby the second model differs from the first only in that it has two additional datasets added to the training regime that are designed to impart simple numerical reasoning strategies of a sort known to improve performance on some of our evaluation datasets but not on others.

Question Answering Semantic Similarity +1

Challenges in Annotating Datasets to Quantify Bias in Under-represented Society

no code implementations11 Sep 2023 Vithya Yogarajan, Gillian Dobbie, Timothy Pistotti, Joshua Bensemann, Kobe Knowles

Recent advances in artificial intelligence, including the development of highly sophisticated large language models (LLM), have proven beneficial in many real-world applications.

Gender Classification

Neuromodulation Gated Transformer

1 code implementation5 May 2023 Kobe Knowles, Joshua Bensemann, Diana Benavides-Prado, Vithya Yogarajan, Michael Witbrock, Gillian Dobbie, Yang Chen

We introduce a novel architecture, the Neuromodulation Gated Transformer (NGT), which is a simple implementation of neuromodulation in transformers via a multiplicative effect.

Input-length-shortening and text generation via attention values

no code implementations14 Mar 2023 Neşet Özkan Tan, Alex Yuxuan Peng, Joshua Bensemann, Qiming Bao, Tim Hartill, Mark Gahegan, Michael Witbrock

Because of the attention mechanism's high computational cost, transformer models usually have an input-length limitation caused by hardware constraints.

Conditional Text Generation text-classification +1

AbductionRules: Training Transformers to Explain Unexpected Inputs

1 code implementation Findings (ACL) 2022 Nathan Young, Qiming Bao, Joshua Bensemann, Michael Witbrock

Transformers have recently been shown to be capable of reliably performing logical reasoning over facts and rules expressed in natural language, but abductive reasoning - inference to the best explanation of an unexpected observation - has been underexplored despite significant applications to scientific discovery, common-sense reasoning, and model interpretability.

Common Sense Reasoning Logical Reasoning

Relating Blindsight and AI: A Review

no code implementations9 Dec 2021 Joshua Bensemann, Qiming Bao, Gaël Gendron, Tim Hartill, Michael Witbrock

If we assume that artificial networks have no form of visual experience, then deficits caused by blindsight give us insights into the processes occurring within visual experience that we can incorporate into artificial neural networks.

Cannot find the paper you are looking for? You can Submit a new open access paper.