๐ฎ
๐ฎ
The Ethereal
Complexity of Training ReLU Neural Network
September 27, 2018 ยท The Ethereal ยท ๐ Discrete Optimization
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Digvijay Boob, Santanu S. Dey, Guanghui Lan
arXiv ID
1809.10787
Category
cs.CC: Computational Complexity
Cross-listed
cs.LG
Citations
79
Venue
Discrete Optimization
Last Checked
1 month ago
Abstract
In this paper, we explore some basic questions on the complexity of training neural networks with ReLU activation function. We show that it is NP-hard to train a two-hidden layer feedforward ReLU neural network. If dimension of the input data and the network topology is fixed, then we show that there exists a polynomial time algorithm for the same training problem. We also show that if sufficient over-parameterization is provided in the first hidden layer of ReLU neural network, then there is a polynomial time algorithm which finds weights such that output of the over-parameterized ReLU neural network matches with the output of the given data.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computational Complexity
๐ฎ
๐ฎ
The Ethereal
An Exponential Separation Between Randomized and Deterministic Complexity in the LOCAL Model
๐ฎ
๐ฎ
The Ethereal
The Parallelism Tradeoff: Limitations of Log-Precision Transformers
๐ฎ
๐ฎ
The Ethereal
The Hardness of Approximation of Euclidean k-means
๐ฎ
๐ฎ
The Ethereal
Slightly Superexponential Parameterized Problems
๐ฎ
๐ฎ
The Ethereal