no code implementations • CMCL (ACL) 2022 • Joshua Bensemann, Alex Peng, Diana Prado, Yang Chen, Neset Tan, Paul Michael Corballis, Patricia Riddle, Michael Witbrock
Attention describes cognitive processes that are important to many human phenomena including reading.
no code implementations • 21 Nov 2023 • Tim Hartill, Joshua Bensemann, Michael Witbrock, Patricia J. Riddle
We train two Language Models in a multitask fashion whereby the second model differs from the first only in that it has two additional datasets added to the training regime that are designed to impart simple numerical reasoning strategies of a sort known to improve performance on some of our evaluation datasets but not on others.
no code implementations • 11 Sep 2023 • Vithya Yogarajan, Gillian Dobbie, Timothy Pistotti, Joshua Bensemann, Kobe Knowles
Recent advances in artificial intelligence, including the development of highly sophisticated large language models (LLM), have proven beneficial in many real-world applications.
1 code implementation • 5 May 2023 • Kobe Knowles, Joshua Bensemann, Diana Benavides-Prado, Vithya Yogarajan, Michael Witbrock, Gillian Dobbie, Yang Chen
We introduce a novel architecture, the Neuromodulation Gated Transformer (NGT), which is a simple implementation of neuromodulation in transformers via a multiplicative effect.
no code implementations • 14 Mar 2023 • Neşet Özkan Tan, Alex Yuxuan Peng, Joshua Bensemann, Qiming Bao, Tim Hartill, Mark Gahegan, Michael Witbrock
Because of the attention mechanism's high computational cost, transformer models usually have an input-length limitation caused by hardware constraints.
1 code implementation • Findings (ACL) 2022 • Nathan Young, Qiming Bao, Joshua Bensemann, Michael Witbrock
Transformers have recently been shown to be capable of reliably performing logical reasoning over facts and rules expressed in natural language, but abductive reasoning - inference to the best explanation of an unexpected observation - has been underexplored despite significant applications to scientific discovery, common-sense reasoning, and model interpretability.
no code implementations • 9 Dec 2021 • Joshua Bensemann, Qiming Bao, Gaël Gendron, Tim Hartill, Michael Witbrock
If we assume that artificial networks have no form of visual experience, then deficits caused by blindsight give us insights into the processes occurring within visual experience that we can incorporate into artificial neural networks.