no code implementations • 7 Dec 2022 • Promit Chakroborty, Somayajulu L. N. Dhulipala, Yifeng Che, Wen Jiang, Benjamin W. Spencer, Jason D. Hales, Michael D. Shields
The multi-fidelity surrogate is assembled by first applying a Gaussian process correction to each low-fidelity model and assigning a model probability based on the model's local predictive accuracy and cost.
1 code implementation • 19 Sep 2022 • Somayajulu L. N. Dhulipala, Yifeng Che, Michael D. Shields
We propose the use of HNNs for performing Bayesian inference efficiently without requiring numerous posterior gradients.
1 code implementation • 12 Aug 2022 • Somayajulu L. N. Dhulipala, Yifeng Che, Michael D. Shields
Compared to traditional NUTS, L-HNNs in NUTS with online error monitoring required 1--2 orders of magnitude fewer numerical gradients of the target density and improved the effective sample size (ESS) per gradient by an order of magnitude.
no code implementations • 6 Jan 2022 • Somayajulu L. N. Dhulipala, Michael D. Shields, Promit Chakroborty, Wen Jiang, Benjamin W. Spencer, Jason D. Hales, Vincent M. Laboure, Zachary M. Prince, Chandrakanth Bolisetti, Yifeng Che
However, TRISO failure probabilities are small and the associated computational models are expensive.
no code implementations • 17 Apr 2021 • Yifeng Che, Joseph Yurko, Koroush Shirvan
Such one-way coupling is result of the high cost induced by the full-core fuel performance analysis, which provides more realistic and accurate prediction of the core-wide response than the "peak rod" analysis.