Universal discretization and sparse sampling recovery
January 14, 2023 ยท Declared Dead ยท ๐ arXiv.org
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
F. Dai, V. Temlyakov
arXiv ID
2301.05962
Category
math.NA: Numerical Analysis
Cross-listed
cs.IT,
math.CA,
math.FA
Citations
15
Venue
arXiv.org
Last Checked
1 month ago
Abstract
Recently, it was discovered that for a given function class $\mathbf{F}$ the error of best linear recovery in the square norm can be bounded above by the Kolmogorov width of $\mathbf{F}$ in the uniform norm. That analysis is based on deep results in discretization of the square norm of functions from finite dimensional subspaces. In this paper we show how very recent results on universal discretization of the square norm of functions from a collection of finite dimensional subspaces lead to an inequality between optimal sparse recovery in the square norm and best sparse approximations in the uniform norm with respect to appropriate dictionaries.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Numerical Analysis
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations
R.I.P.
๐ป
Ghosted
PDE-Net: Learning PDEs from Data
R.I.P.
๐ป
Ghosted
Efficient tensor completion for color image and video recovery: Low-rank tensor train
R.I.P.
๐ป
Ghosted
Tensor Ring Decomposition
R.I.P.
๐ป
Ghosted
Machine learning approximation algorithms for high-dimensional fully nonlinear partial differential equations and second-order backward stochastic differential equations
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted