no code implementations • 30 Oct 2024 • Peter Belcak, Roger Wattenhofer
It is staggering that words of the English language, which are on average represented by 5--6 bytes of ASCII, require as much as 24 kilobytes when served to large language models.
3 code implementations • 15 Nov 2023 • Peter Belcak, Roger Wattenhofer
Language models only really need to use an exponential fraction of their neurons for individual inferences.
4 code implementations • 28 Aug 2023 • Peter Belcak, Roger Wattenhofer
We break the linear link between the layer size and its inference cost by introducing the fast feedforward (FFF) architecture, a log-time alternative to feedforward networks.
no code implementations • 31 May 2023 • Peter Belcak, Luca A. Lanzendörfer, Roger Wattenhofer
We conduct a preliminary inquiry into the ability of generative transformer models to deductively reason from premises provided.
1 code implementation • 29 Oct 2022 • Peter Belcak, Roger Wattenhofer
We propose a novel, fully explainable neural approach to synthesis of combinatorial logic circuits from input-output examples.
no code implementations • 22 Aug 2022 • Peter Belcak, Roger Wattenhofer
These programs characterise linear long-distance relationships between the given two vertex sets in the context of the whole graph.
no code implementations • 15 Oct 2020 • Peter Belcak
The LL(finite) parsing strategy for parsing of LL(k) grammars where k needs not to be known is presented.
no code implementations • 18 Aug 2020 • Peter Belcak, Jan-Peter Calliess, Stefan Zohren
As a simple illustration, we employ our toolbox to investigate the role of the order processing delay in normal trading and for the scenario of a significant price change.