1 code implementation • 22 Feb 2025 • H. V. AlquBoj, Hilal AlQuabeh, Velibor Bojkovic, Tatsuya Hiraoka, Ahmed Oumar El-Shangiti, Munachiso Nwadike, Kentaro Inui
Humans are believed to perceive numbers on a logarithmic mental number line, where smaller values are represented with greater resolution than larger ones.
no code implementations • 17 Oct 2024 • Ahmed Oumar El-Shangiti, Tatsuya Hiraoka, Hilal AlQuabeh, Benjamin Heinzerling, Kentaro Inui
This paper investigates whether large language models (LLMs) utilize numerical attributes encoded in a low-dimensional subspace of the embedding space when answering logical comparison questions (e. g., Was Cristiano born before Messi?).
no code implementations • 2 Feb 2024 • Hilal AlQuabeh, William de Vazelhes, Bin Gu
Recently, an OGD algorithm emerged, employing gradient computation involving prior and most recent examples, a step that effectively reduces algorithmic complexity to $O(T)$, with $T$ being the number of received examples.
1 code implementation • 10 Oct 2023 • Hilal AlQuabeh, Bhaskar Mukhoty, Bin Gu
Specifically, we establish a clear connection between the variance of online gradients and the regret, and construct online gradients using the most recent stratified samples with a limited buffer of size of $s$ representing all past data, which have a complexity of $O(sT)$ and employs $O(\sqrt{T}\log{T})$ random Fourier features for kernel approximation.
no code implementations • 29 Sep 2022 • Hilal AlQuabeh, Farha AlBreiki, Dilshod Azizov
One of these approaches is reducing the gradient variance through adaptive sampling to solve large-scale optimization's empirical risk minimization (ERM) problems.
no code implementations • 8 Aug 2022 • Hilal AlQuabeh, Aliakbar Abdurahimov
Recent research has, however, offered adaptive sample size training for smooth loss functions as a better strategy in terms of convergence and complexity, but without a comprehensive theoretical study.
no code implementations • 17 Aug 2021 • Hilal AlQuabeh, Ameera Bawazeer, Abdulateef Alhashmi
Results show that the pretext process in the self-supervised learning improves the accuracy around 15% in the downstream classification task when compared to the plain supervised learning.