no code implementations • 22 Mar 2024 • John Fischer, Marko Orescanin, Justin Loomis, Patrick McClure
Aggregation strategies have been developed to pool or fuse the weights and biases of distributed deterministic models; however, modern deterministic deep learning (DL) models are often poorly calibrated and lack the ability to communicate a measure of epistemic uncertainty in prediction, which is desirable for remote sensing platforms and safety-critical applications.
1 code implementation • 10 Jan 2024 • John Fischer, Marko Orescanin, Eric Eckstrand
We demonstrate, for the first time, that it is possible to transfer calibrated uncertainty information along with knowledge from upstream tasks to enhance a model's capability to perform downstream tasks.