1 code implementation • 1 Feb 2024 • Toni J. B. Liu, Nicolas Boullé, Raphaël Sarfati, Christopher J. Earls
Pretrained large language models (LLMs) are surprisingly effective at performing zero-shot tasks, including time-series forecasting.
1 code implementation • NeurIPS 2023 • Albert Tseng, Tao Yu, Toni J. B. Liu, Christopher De Sa
These networks rely heavily on the dot product attention operator, which computes the similarity between two points by taking their inner product.
1 code implementation • 24 May 2023 • Tao Yu, Toni J. B. Liu, Albert Tseng, Christopher De Sa
Specifically, we model partial orders as subset relations between shadows formed by a light source and opaque objects in hyperbolic space.