Resum
How to structure the topology of a neural net (the hyperparameters optimization) is a recurring problem of crucial importance for both the quality and rapidity of the learning process, which will be then translated onto the final outputs. In this article, the authors investigate a smart two-step procedure that formalizes the application of deep feed-forward neural nets in the problem of the calibration of the SABR option-pricing model. The analysis is performed without the need of manually preparing the network topology, that is instead optimally chosen by means of a Bayesian algorithm. An extensive numerical experiment shows that their approach possesses superior approximation, calibration and retrieval properties when compared to the Hagan’s formula and the ZC map.
| Idioma original | Anglès |
|---|---|
| Pàgines (de-a) | 147-181 |
| Nombre de pàgines | 35 |
| Revista | Journal of Financial Data Science |
| Volum | 6 |
| Número | 3 |
| DOIs | |
| Estat de la publicació | Publicada - 1 de juny 2024 |
Fingerprint
Navegar pels temes de recerca de 'Smart Deep Learning Calibration of the SABR Model'. Junts formen un fingerprint únic.Com citar-ho
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver