Sparsest Error Detection via Sparsity Invariant Transformation based $\ell_1$ Minimization
May 20, 2015 Β· Declared Dead Β· π arXiv.org
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Suzhen Wang, Sheng Han, Zhiguo Zhang, Wing Shing Wong
arXiv ID
1505.05290
Category
stat.ME
Cross-listed
cs.IT
Citations
0
Venue
arXiv.org
Last Checked
2 months ago
Abstract
This paper presents a new method, referred to here as the sparsity invariant transformation based $\ell_1$ minimization, to solve the $\ell_0$ minimization problem for an over-determined linear system corrupted by additive sparse errors with arbitrary intensity. Many previous works have shown that $\ell_1$ minimization can be applied to realize sparse error detection in many over-determined linear systems. However, performance of this approach is strongly dependent on the structure of the measurement matrix, which limits application possibility in practical problems. Here, we present a new approach based on transforming the $\ell_0$ minimization problem by a linear transformation that keeps sparsest solutions invariant. We call such a property a sparsity invariant property (SIP), and a linear transformation with SIP is referred to as a sparsity invariant transformation (SIT). We propose the SIT-based $\ell_1$ minimization method by using an SIT in conjunction with $\ell_1$ relaxation on the $\ell_0$ minimization problem. We prove that for any over-determined linear system, there always exists a specific class of SIT's that guarantees a solution to the SIT-based $\ell_1$ minimization is a sparsest-errors solution. Besides, a randomized algorithm based on Monte Carlo simulation is proposed to search for a feasible SIT.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β stat.ME
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Performance Metrics (Error Measures) in Machine Learning Regression, Forecasting and Prognostics: Properties and Typology
R.I.P.
π»
Ghosted
External Validity: From Do-Calculus to Transportability Across Populations
R.I.P.
π»
Ghosted
Least Ambiguous Set-Valued Classifiers with Bounded Error Levels
R.I.P.
π»
Ghosted
Doubly Robust Policy Evaluation and Optimization
R.I.P.
π»
Ghosted
Comparison of Bayesian predictive methods for model selection
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted