๐ฎ
๐ฎ
The Ethereal
Lower bounds over Boolean inputs for deep neural networks with ReLU gates
November 08, 2017 ยท The Ethereal ยท ๐ Electron. Colloquium Comput. Complex.
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Anirbit Mukherjee, Amitabh Basu
arXiv ID
1711.03073
Category
cs.CC: Computational Complexity
Cross-listed
cs.LG,
cs.NE,
stat.ML
Citations
23
Venue
Electron. Colloquium Comput. Complex.
Last Checked
2 months ago
Abstract
Motivated by the resurgence of neural networks in being able to solve complex learning tasks we undertake a study of high depth networks using ReLU gates which implement the function $x \mapsto \max\{0,x\}$. We try to understand the role of depth in such neural networks by showing size lowerbounds against such network architectures in parameter regimes hitherto unexplored. In particular we show the following two main results about neural nets computing Boolean functions of input dimension $n$, 1. We use the method of random restrictions to show almost linear, $ฮฉ(ฮต^{2(1-ฮด)}n^{1-ฮด})$, lower bound for completely weight unrestricted LTF-of-ReLU circuits to match the Andreev function on at least $\frac{1}{2} +ฮต$ fraction of the inputs for $ฮต> \sqrt{2\frac{\log^{\frac {2}{2-ฮด}}(n)}{n}}$ for any $ฮด\in (0,\frac 1 2)$ 2. We use the method of sign-rank to show exponential in dimension lower bounds for ReLU circuits ending in a LTF gate and of depths upto $O(n^ฮพ)$ with $ฮพ< \frac{1}{8}$ with some restrictions on the weights in the bottom most layer. All other weights in these circuits are kept unrestricted. This in turns also implies the same lowerbounds for LTF circuits with the same architecture and the same weight restrictions on their bottom most layer. Along the way we also show that there exists a $\mathbb{R}^ n\rightarrow \mathbb{R}$ Sum-of-ReLU-of-ReLU function which Sum-of-ReLU neural nets can never represent no matter how large they are allowed to be.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computational Complexity
๐ฎ
๐ฎ
The Ethereal
An Exponential Separation Between Randomized and Deterministic Complexity in the LOCAL Model
๐ฎ
๐ฎ
The Ethereal
The Parallelism Tradeoff: Limitations of Log-Precision Transformers
๐ฎ
๐ฎ
The Ethereal
The Hardness of Approximation of Euclidean k-means
๐ฎ
๐ฎ
The Ethereal
Slightly Superexponential Parameterized Problems
๐ฎ
๐ฎ
The Ethereal