1 code implementation • 4 Jan 2024 • Parvin Malekzadeh, Konstantinos N. Plataniotis, Zissis Poulos, Zeyu Wang
Distributional Reinforcement Learning (RL) estimates return distribution mainly by learning quantile values via minimizing the quantile Huber loss function, entailing a threshold parameter often selected heuristically or via hyperparameter search, which may not generalize well and can be suboptimal.
1 code implementation • 10 May 2022 • Jay Cao, Jacky Chen, Soroush Farghadani, John Hull, Zissis Poulos, Zeyu Wang, Jun Yuan
We show how D4PG can be used in conjunction with quantile regression to develop a hedging strategy for a trader responsible for derivatives that arrive stochastically and depend on a single underlying asset.
no code implementations • 29 Mar 2021 • Jay Cao, Jacky Chen, John Hull, Zissis Poulos
This paper shows how reinforcement learning can be used to derive optimal hedging strategies for derivatives when there are transaction costs.
no code implementations • 22 Mar 2021 • Jay Cao, Jacky Chen, John Hull, Zissis Poulos
We refer to this as the model calibration approach (MCA).
no code implementations • 7 Feb 2021 • Maxime Bergeron, Nicholas Fung, John Hull, Zissis Poulos
As a dividend of our first step, the synthetic surfaces produced can also be used in stress testing, in market simulators for developing quantitative investment strategies, and for the valuation of exotic options.
no code implementations • 15 Oct 2019 • Zissis Poulos, Ali Nouri, Andreas Moshovos
We reduce training time in convolutional networks (CNNs) with a method that, for some of the mini-batches: a) scales down the resolution of input images via downsampling, and b) reduces the forward pass operations via pooling on the convolution filters.
no code implementations • 9 Mar 2018 • Alberto Delmas, Patrick Judd, Dylan Malone Stuart, Zissis Poulos, Mostafa Mahmoud, Sayeh Sharify, Milos Nikolic, Andreas Moshovos
We show that, during inference with Convolutional Neural Networks (CNNs), more than 2x to $8x ineffectual work can be exposed if instead of targeting those weights and activations that are zero, we target different combinations of value stream properties.