no code implementations • 17 Apr 2025 • Nischal Mainali, Lucas Teixeira
Transformer models exhibit remarkable in-context learning (ICL), adapting to novel tasks from examples within their context, yet the underlying mechanisms remain largely mysterious.
no code implementations • 24 May 2024 • Adam S. Shai, Sarah E. Marzen, Lucas Teixeira, Alexander Gietelink Oldenziel, Paul M. Riechers
What computational structure are we building into large language models when we train them on next-token prediction?
2 code implementations • 17 Jan 2020 • Lucas Teixeira, Martin R. Oswald, Marc Pollefeys, Margarita Chli
In this paper, we propose a depth completion and uncertainty estimation approach that better handles the challenges of aerial platforms, such as large viewpoint and depth variations, and limited computing resources.