Cosmology with Persistent Homology: Parameter Inference via Machine Learning
December 19, 2024 Β· Declared Dead Β· π Journal of Cosmology and Astroparticle Physics
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Juan Calles, Jacky H. T. Yip, Gabriella Contardo, Jorge NoreΓ±a, Adam Rouhiainen, Gary Shiu
arXiv ID
2412.15405
Category
astro-ph.CO
Cross-listed
cs.LG,
math.AT
Citations
2
Venue
Journal of Cosmology and Astroparticle Physics
Last Checked
1 month ago
Abstract
Building upon [2308.02636], we investigate the constraining power of persistent homology on cosmological parameters and primordial non-Gaussianity in a likelihood-free inference pipeline utilizing machine learning. We evaluate the ability of Persistence Images (PIs) to infer parameters, comparing them to the combined Power Spectrum and Bispectrum (PS/BS). We also compare two classes of models: neural-based and tree-based. PIs consistently lead to better predictions compared to the combined PS/BS for parameters that can be constrained, i.e., for $\{Ξ©_{\rm m}, Ο_8, n_{\rm s}, f_{\rm NL}^{\rm loc}\}$. PIs perform particularly well for $f_{\rm NL}^{\rm loc}$, highlighting the potential of persistent homology for constraining primordial non-Gaussianity. Our results indicate that combining PIs with PS/BS provides only marginal gains, indicating that the PS/BS contains little additional or complementary information to the PIs. Finally, we provide a visualization of the most important topological features for $f_{\rm NL}^{\rm loc}$ and for $Ξ©_{\rm m}$. This reveals that clusters and voids (0-cycles and 2-cycles) are most informative for $Ξ©_{\rm m}$, while $f_{\rm NL}^{\rm loc}$ is additionally informed by filaments (1-cycles).
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β astro-ph.CO
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Learning to Predict the Cosmological Structure Formation
R.I.P.
π»
Ghosted
Estimating Cosmological Parameters from the Dark Matter Distribution
R.I.P.
π»
Ghosted
DeepCMB: Lensing Reconstruction of the Cosmic Microwave Background with Deep Neural Networks
R.I.P.
π»
Ghosted
Information theory for fields
R.I.P.
π»
Ghosted
Exhaustive Symbolic Regression
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted